[ 466.756348] env[62914]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62914) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 466.756709] env[62914]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62914) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 466.756826] env[62914]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62914) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 466.757124] env[62914]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 466.849749] env[62914]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62914) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 466.859797] env[62914]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62914) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 467.459691] env[62914]: INFO nova.virt.driver [None req-89d542e0-6486-40be-9cee-df03915e7f13 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 467.528711] env[62914]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 467.528872] env[62914]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 467.528961] env[62914]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62914) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 470.695878] env[62914]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-bb7973da-cd1a-4cf9-a38b-dc05e8de6a52 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.711634] env[62914]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62914) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 470.711819] env[62914]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-98eaf480-b246-4ab0-8651-c03b5d97d4cb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.746517] env[62914]: INFO oslo_vmware.api [-] Successfully established new session; session ID is f9a6f. [ 470.746731] env[62914]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.218s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 470.747213] env[62914]: INFO nova.virt.vmwareapi.driver [None req-89d542e0-6486-40be-9cee-df03915e7f13 None None] VMware vCenter version: 7.0.3 [ 470.750579] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf527b1-0da1-4ab9-9deb-30301347151d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.767660] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98569426-55c4-4146-a0fd-f0284773693a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.773600] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c69811-5078-47d3-88b0-e525ccd1e000 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.779893] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305dd2e7-2507-4aa2-89dd-b5e523048046 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.792588] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c1af82-bbc8-4f74-8d01-bca4b8619de5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.798510] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ff505a3-258d-478e-baef-8eae5359884d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.828710] env[62914]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-8dcef3c2-bcee-4b30-916a-42629b6a4889 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.833988] env[62914]: DEBUG nova.virt.vmwareapi.driver [None req-89d542e0-6486-40be-9cee-df03915e7f13 None None] Extension org.openstack.compute already exists. {{(pid=62914) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 470.836655] env[62914]: INFO nova.compute.provider_config [None req-89d542e0-6486-40be-9cee-df03915e7f13 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 471.340512] env[62914]: DEBUG nova.context [None req-89d542e0-6486-40be-9cee-df03915e7f13 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),99cfa7b5-f243-4608-a55d-4e73fcc2bbf2(cell1) {{(pid=62914) load_cells /opt/stack/nova/nova/context.py:464}} [ 471.342596] env[62914]: DEBUG oslo_concurrency.lockutils [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 471.342822] env[62914]: DEBUG oslo_concurrency.lockutils [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 471.343506] env[62914]: DEBUG oslo_concurrency.lockutils [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 471.343926] env[62914]: DEBUG oslo_concurrency.lockutils [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Acquiring lock "99cfa7b5-f243-4608-a55d-4e73fcc2bbf2" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 471.344166] env[62914]: DEBUG oslo_concurrency.lockutils [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Lock "99cfa7b5-f243-4608-a55d-4e73fcc2bbf2" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 471.345193] env[62914]: DEBUG oslo_concurrency.lockutils [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Lock "99cfa7b5-f243-4608-a55d-4e73fcc2bbf2" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 471.364828] env[62914]: INFO dbcounter [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Registered counter for database nova_cell0 [ 471.372934] env[62914]: INFO dbcounter [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Registered counter for database nova_cell1 [ 471.376449] env[62914]: DEBUG oslo_db.sqlalchemy.engines [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62914) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 471.376814] env[62914]: DEBUG oslo_db.sqlalchemy.engines [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62914) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 471.381713] env[62914]: ERROR nova.db.main.api [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 471.381713] env[62914]: result = function(*args, **kwargs) [ 471.381713] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 471.381713] env[62914]: return func(*args, **kwargs) [ 471.381713] env[62914]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 471.381713] env[62914]: result = fn(*args, **kwargs) [ 471.381713] env[62914]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 471.381713] env[62914]: return f(*args, **kwargs) [ 471.381713] env[62914]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 471.381713] env[62914]: return db.service_get_minimum_version(context, binaries) [ 471.381713] env[62914]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 471.381713] env[62914]: _check_db_access() [ 471.381713] env[62914]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 471.381713] env[62914]: stacktrace = ''.join(traceback.format_stack()) [ 471.381713] env[62914]: [ 471.382422] env[62914]: ERROR nova.db.main.api [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 471.382422] env[62914]: result = function(*args, **kwargs) [ 471.382422] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 471.382422] env[62914]: return func(*args, **kwargs) [ 471.382422] env[62914]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 471.382422] env[62914]: result = fn(*args, **kwargs) [ 471.382422] env[62914]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 471.382422] env[62914]: return f(*args, **kwargs) [ 471.382422] env[62914]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 471.382422] env[62914]: return db.service_get_minimum_version(context, binaries) [ 471.382422] env[62914]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 471.382422] env[62914]: _check_db_access() [ 471.382422] env[62914]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 471.382422] env[62914]: stacktrace = ''.join(traceback.format_stack()) [ 471.382422] env[62914]: [ 471.382798] env[62914]: WARNING nova.objects.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 471.382933] env[62914]: WARNING nova.objects.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Failed to get minimum service version for cell 99cfa7b5-f243-4608-a55d-4e73fcc2bbf2 [ 471.383347] env[62914]: DEBUG oslo_concurrency.lockutils [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Acquiring lock "singleton_lock" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 471.383509] env[62914]: DEBUG oslo_concurrency.lockutils [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Acquired lock "singleton_lock" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 471.383751] env[62914]: DEBUG oslo_concurrency.lockutils [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Releasing lock "singleton_lock" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 471.384071] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Full set of CONF: {{(pid=62914) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 471.384249] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ******************************************************************************** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 471.384384] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Configuration options gathered from: {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 471.384523] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 471.384718] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 471.384844] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ================================================================================ {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 471.385072] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] allow_resize_to_same_host = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.385288] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] arq_binding_timeout = 300 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.385430] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] backdoor_port = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.385558] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] backdoor_socket = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.385723] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] block_device_allocate_retries = 60 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.385881] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] block_device_allocate_retries_interval = 3 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.386062] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cert = self.pem {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.386232] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.386402] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute_monitors = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.386567] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] config_dir = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.386735] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] config_drive_format = iso9660 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.386866] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.387040] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] config_source = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.387213] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] console_host = devstack {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.387378] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] control_exchange = nova {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.387536] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cpu_allocation_ratio = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.387698] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] daemon = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.387865] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] debug = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.388120] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] default_access_ip_network_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.388319] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] default_availability_zone = nova {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.388483] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] default_ephemeral_format = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.388646] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] default_green_pool_size = 1000 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.388889] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.389074] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] default_schedule_zone = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.389238] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] disk_allocation_ratio = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.389404] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] enable_new_services = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.389583] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] enabled_apis = ['osapi_compute'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.389752] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] enabled_ssl_apis = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.389914] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] flat_injected = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.390084] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] force_config_drive = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.390248] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] force_raw_images = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.390419] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] graceful_shutdown_timeout = 5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.390578] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] heal_instance_info_cache_interval = 60 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.390794] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] host = cpu-1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.390972] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.391151] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.391318] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.391530] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.391694] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] instance_build_timeout = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.391855] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] instance_delete_interval = 300 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.392033] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] instance_format = [instance: %(uuid)s] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.392208] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] instance_name_template = instance-%08x {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.392371] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] instance_usage_audit = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.392540] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] instance_usage_audit_period = month {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.392707] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.392870] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.393047] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] internal_service_availability_zone = internal {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.393210] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] key = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.393373] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] live_migration_retry_count = 30 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.393541] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] log_color = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.393706] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] log_config_append = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.393871] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.394043] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] log_dir = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.394237] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] log_file = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.394376] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] log_options = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.394545] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] log_rotate_interval = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.394718] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] log_rotate_interval_type = days {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.394886] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] log_rotation_type = none {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.395026] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.395157] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.395354] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.395527] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.395655] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.395822] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] long_rpc_timeout = 1800 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.395983] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] max_concurrent_builds = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.396160] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] max_concurrent_live_migrations = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.396323] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] max_concurrent_snapshots = 5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.396483] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] max_local_block_devices = 3 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.396640] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] max_logfile_count = 30 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.396797] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] max_logfile_size_mb = 200 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.396955] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] maximum_instance_delete_attempts = 5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.397138] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] metadata_listen = 0.0.0.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.397309] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] metadata_listen_port = 8775 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.397476] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] metadata_workers = 2 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.397637] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] migrate_max_retries = -1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.397802] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] mkisofs_cmd = genisoimage {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.398016] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.398158] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] my_ip = 10.180.1.21 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.398324] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] network_allocate_retries = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.398500] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.398667] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.398828] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] osapi_compute_listen_port = 8774 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.398995] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] osapi_compute_unique_server_name_scope = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.399175] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] osapi_compute_workers = 2 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.399340] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] password_length = 12 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.399498] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] periodic_enable = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.399653] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] periodic_fuzzy_delay = 60 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.399818] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] pointer_model = usbtablet {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.399981] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] preallocate_images = none {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.400151] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] publish_errors = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.400281] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] pybasedir = /opt/stack/nova {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.400435] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ram_allocation_ratio = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.400590] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] rate_limit_burst = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.400757] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] rate_limit_except_level = CRITICAL {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.400915] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] rate_limit_interval = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.401086] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] reboot_timeout = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.401250] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] reclaim_instance_interval = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.401407] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] record = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.401572] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] reimage_timeout_per_gb = 60 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.401735] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] report_interval = 120 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.401892] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] rescue_timeout = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.402058] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] reserved_host_cpus = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.402217] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] reserved_host_disk_mb = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.402374] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] reserved_host_memory_mb = 512 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.402532] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] reserved_huge_pages = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.402687] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] resize_confirm_window = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.402843] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] resize_fs_using_block_device = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.402997] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] resume_guests_state_on_host_boot = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.403176] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.403338] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] rpc_response_timeout = 60 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.403494] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] run_external_periodic_tasks = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.403656] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] running_deleted_instance_action = reap {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.403813] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.403969] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] running_deleted_instance_timeout = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.404162] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] scheduler_instance_sync_interval = 120 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.404346] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] service_down_time = 720 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.404517] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] servicegroup_driver = db {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.404674] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] shell_completion = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.404834] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] shelved_offload_time = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.404991] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] shelved_poll_interval = 3600 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.405175] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] shutdown_timeout = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.405380] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] source_is_ipv6 = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.405547] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ssl_only = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.405792] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.405961] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] sync_power_state_interval = 600 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.406137] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] sync_power_state_pool_size = 1000 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.406307] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] syslog_log_facility = LOG_USER {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.406464] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] tempdir = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.406619] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] timeout_nbd = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.406783] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] transport_url = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.406942] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] update_resources_interval = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.407113] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] use_cow_images = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.407275] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] use_eventlog = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.407430] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] use_journal = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.407584] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] use_json = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.407739] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] use_rootwrap_daemon = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.407890] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] use_stderr = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.408054] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] use_syslog = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.408212] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vcpu_pin_set = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.408377] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vif_plugging_is_fatal = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.408541] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vif_plugging_timeout = 300 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.408701] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] virt_mkfs = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.408858] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] volume_usage_poll_interval = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.409025] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] watch_log_file = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.409197] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] web = /usr/share/spice-html5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.409385] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.409550] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.409713] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.409878] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_concurrency.disable_process_locking = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.410434] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.410631] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.410804] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.410991] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.411185] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.411359] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.411542] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.auth_strategy = keystone {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.411712] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.compute_link_prefix = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.411890] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.412077] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.dhcp_domain = novalocal {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.412254] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.enable_instance_password = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.412422] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.glance_link_prefix = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.412588] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.412761] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.412924] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.instance_list_per_project_cells = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.413099] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.list_records_by_skipping_down_cells = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.413268] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.local_metadata_per_cell = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.413437] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.max_limit = 1000 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.413605] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.metadata_cache_expiration = 15 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.413780] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.neutron_default_tenant_id = default {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.413954] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.response_validation = warn {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.414199] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.use_neutron_default_nets = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.414352] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.414523] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.414692] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.414864] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.415045] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.vendordata_dynamic_targets = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.415231] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.vendordata_jsonfile_path = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.415433] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.415630] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.backend = dogpile.cache.memcached {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.415798] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.backend_argument = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.415971] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.config_prefix = cache.oslo {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.416153] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.dead_timeout = 60.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.416319] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.debug_cache_backend = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.416485] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.enable_retry_client = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.416644] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.enable_socket_keepalive = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.416811] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.enabled = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.416972] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.enforce_fips_mode = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.417151] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.expiration_time = 600 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.417344] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.hashclient_retry_attempts = 2 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.417531] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.417698] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.memcache_dead_retry = 300 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.417859] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.memcache_password = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.418035] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.418205] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.418372] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.memcache_pool_maxsize = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.418534] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.418693] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.memcache_sasl_enabled = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.418871] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.419045] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.419210] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.memcache_username = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.419377] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.proxies = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.419539] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.redis_db = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.419699] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.redis_password = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.419869] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.420055] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.420232] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.redis_server = localhost:6379 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.420400] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.redis_socket_timeout = 1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.420559] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.redis_username = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.420722] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.retry_attempts = 2 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.420886] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.retry_delay = 0.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.421060] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.socket_keepalive_count = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.421229] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.socket_keepalive_idle = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.421393] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.socket_keepalive_interval = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.421551] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.tls_allowed_ciphers = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.421707] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.tls_cafile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.421861] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.tls_certfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.422048] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.tls_enabled = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.422272] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cache.tls_keyfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.422399] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.auth_section = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.422576] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.auth_type = password {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.422738] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.cafile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.422914] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.423086] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.certfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.423253] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.collect_timing = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.423416] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.cross_az_attach = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.423577] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.debug = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.423737] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.endpoint_template = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.423900] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.http_retries = 3 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.424069] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.insecure = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.424310] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.keyfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.424440] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.os_region_name = RegionOne {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.424609] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.split_loggers = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.424771] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cinder.timeout = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.424944] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.425135] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute.cpu_dedicated_set = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.425310] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute.cpu_shared_set = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.425477] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute.image_type_exclude_list = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.425640] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.425802] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.425964] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.426142] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.426315] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.426478] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute.resource_provider_association_refresh = 300 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.426636] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.426798] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute.shutdown_retry_interval = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.426977] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.427171] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] conductor.workers = 2 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.427352] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] console.allowed_origins = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.427513] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] console.ssl_ciphers = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.427682] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] console.ssl_minimum_version = default {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.427850] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] consoleauth.enforce_session_timeout = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.428028] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] consoleauth.token_ttl = 600 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.428207] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.cafile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.428370] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.certfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.428531] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.collect_timing = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.428687] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.connect_retries = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.428842] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.connect_retry_delay = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.428997] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.endpoint_override = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.429174] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.insecure = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.429333] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.keyfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.429490] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.max_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.429646] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.min_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.429802] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.region_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.429958] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.retriable_status_codes = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.430129] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.service_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.430301] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.service_type = accelerator {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.430461] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.split_loggers = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.430616] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.status_code_retries = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.430771] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.status_code_retry_delay = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.430928] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.timeout = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.431121] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.431286] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] cyborg.version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.431465] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.backend = sqlalchemy {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.431635] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.connection = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.431800] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.connection_debug = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.431970] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.connection_parameters = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.432181] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.connection_recycle_time = 3600 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.432352] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.connection_trace = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.432514] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.db_inc_retry_interval = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.432678] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.db_max_retries = 20 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.432839] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.db_max_retry_interval = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.432998] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.db_retry_interval = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.433177] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.max_overflow = 50 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.433342] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.max_pool_size = 5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.433504] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.max_retries = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.433671] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.433831] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.mysql_wsrep_sync_wait = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.433988] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.pool_timeout = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.434194] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.retry_interval = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.434372] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.slave_connection = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.434530] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.sqlite_synchronous = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.434691] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] database.use_db_reconnect = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.434871] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.backend = sqlalchemy {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.435054] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.connection = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.435249] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.connection_debug = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.435428] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.connection_parameters = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.435593] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.connection_recycle_time = 3600 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.435756] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.connection_trace = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.435918] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.db_inc_retry_interval = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.436093] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.db_max_retries = 20 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.436260] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.db_max_retry_interval = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.436426] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.db_retry_interval = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.436584] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.max_overflow = 50 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.436746] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.max_pool_size = 5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.436907] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.max_retries = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.437091] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.437256] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.437415] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.pool_timeout = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.437579] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.retry_interval = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.437738] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.slave_connection = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.437898] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] api_database.sqlite_synchronous = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.438083] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] devices.enabled_mdev_types = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.438266] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.438442] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.438605] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ephemeral_storage_encryption.enabled = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.438769] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.438938] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.api_servers = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.439114] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.cafile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.439280] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.certfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.439440] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.collect_timing = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.439597] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.connect_retries = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.439753] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.connect_retry_delay = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.439914] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.debug = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.440091] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.default_trusted_certificate_ids = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.440259] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.enable_certificate_validation = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.440425] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.enable_rbd_download = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.440582] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.endpoint_override = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.440745] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.insecure = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.440907] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.keyfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.441076] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.max_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.441238] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.min_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.441403] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.num_retries = 3 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.441572] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.rbd_ceph_conf = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.441733] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.rbd_connect_timeout = 5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.441902] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.rbd_pool = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.442125] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.rbd_user = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.442285] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.region_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.442448] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.retriable_status_codes = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.442607] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.service_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.442779] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.service_type = image {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.442941] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.split_loggers = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.443120] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.status_code_retries = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.443284] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.status_code_retry_delay = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.443443] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.timeout = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.443621] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.443788] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.verify_glance_signatures = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.443944] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] glance.version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.444156] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] guestfs.debug = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.444327] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] mks.enabled = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.444688] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.444877] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] image_cache.manager_interval = 2400 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.445060] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] image_cache.precache_concurrency = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.445264] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] image_cache.remove_unused_base_images = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.445443] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.445610] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.445790] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] image_cache.subdirectory_name = _base {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.445960] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.api_max_retries = 60 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.446147] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.api_retry_interval = 2 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.446306] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.auth_section = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.446470] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.auth_type = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.446627] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.cafile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.446783] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.certfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.446946] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.collect_timing = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.447123] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.conductor_group = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.447285] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.connect_retries = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.447448] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.connect_retry_delay = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.447597] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.endpoint_override = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.447763] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.insecure = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.447914] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.keyfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.448083] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.max_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.448245] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.min_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.448412] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.peer_list = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.448572] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.region_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.448729] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.retriable_status_codes = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.448892] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.serial_console_state_timeout = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.449063] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.service_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.449237] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.service_type = baremetal {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.449400] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.shard = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.449561] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.split_loggers = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.449721] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.status_code_retries = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.449879] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.status_code_retry_delay = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.450046] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.timeout = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.450233] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.450393] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ironic.version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.450615] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.450808] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] key_manager.fixed_key = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.450994] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.451172] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.barbican_api_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.451336] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.barbican_endpoint = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.451507] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.barbican_endpoint_type = public {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.451668] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.barbican_region_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.451827] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.cafile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.452065] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.certfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.452198] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.collect_timing = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.452366] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.insecure = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.452528] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.keyfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.452692] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.number_of_retries = 60 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.452853] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.retry_delay = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.453025] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.send_service_user_token = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.453192] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.split_loggers = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.453352] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.timeout = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.453512] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.verify_ssl = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.453667] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican.verify_ssl_path = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.453831] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican_service_user.auth_section = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.453990] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican_service_user.auth_type = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.454257] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican_service_user.cafile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.454349] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican_service_user.certfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.454504] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican_service_user.collect_timing = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.454704] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican_service_user.insecure = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.454819] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican_service_user.keyfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.454978] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican_service_user.split_loggers = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.455181] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] barbican_service_user.timeout = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.455362] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vault.approle_role_id = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.455521] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vault.approle_secret_id = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.455690] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vault.kv_mountpoint = secret {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.455849] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vault.kv_path = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.456024] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vault.kv_version = 2 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.456193] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vault.namespace = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.456352] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vault.root_token_id = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.456512] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vault.ssl_ca_crt_file = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.456676] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vault.timeout = 60.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.456836] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vault.use_ssl = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.457021] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.457191] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.auth_section = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.457354] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.auth_type = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.457512] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.cafile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.457669] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.certfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.457831] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.collect_timing = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.457988] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.connect_retries = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.458170] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.connect_retry_delay = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.458324] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.endpoint_override = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.458483] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.insecure = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.458637] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.keyfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.458791] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.max_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.458945] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.min_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.459114] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.region_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.459273] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.retriable_status_codes = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.459432] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.service_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.459596] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.service_type = identity {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.459754] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.split_loggers = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.459909] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.status_code_retries = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.460077] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.status_code_retry_delay = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.460235] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.timeout = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.460415] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.460572] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] keystone.version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.460771] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.connection_uri = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.460930] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.cpu_mode = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.461107] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.461281] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.cpu_models = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.461451] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.cpu_power_governor_high = performance {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.461618] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.461780] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.cpu_power_management = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.461947] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.462156] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.device_detach_attempts = 8 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.462332] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.device_detach_timeout = 20 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.462499] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.disk_cachemodes = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.462659] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.disk_prefix = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.462822] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.enabled_perf_events = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.462983] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.file_backed_memory = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.463163] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.gid_maps = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.463325] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.hw_disk_discard = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.463481] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.hw_machine_type = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.463653] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.images_rbd_ceph_conf = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.463819] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.463982] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.464190] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.images_rbd_glance_store_name = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.464377] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.images_rbd_pool = rbd {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.464551] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.images_type = default {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.464714] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.images_volume_group = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.464879] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.inject_key = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.465054] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.inject_partition = -2 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.465247] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.inject_password = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.465417] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.iscsi_iface = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.465580] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.iser_use_multipath = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.465745] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.465906] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.466080] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.live_migration_downtime = 500 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.466248] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.466411] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.466573] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.live_migration_inbound_addr = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.466747] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.466914] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.467086] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.live_migration_scheme = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.467274] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.live_migration_timeout_action = abort {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.467443] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.live_migration_tunnelled = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.467604] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.live_migration_uri = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.467767] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.live_migration_with_native_tls = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.467927] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.max_queues = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.468104] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.468365] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.468533] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.nfs_mount_options = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.468832] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.469020] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.469206] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.469383] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.469548] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.469710] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.num_pcie_ports = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.469877] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.470054] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.pmem_namespaces = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.470222] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.quobyte_client_cfg = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.470514] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.470691] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.470857] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.471032] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.471201] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.rbd_secret_uuid = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.471363] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.rbd_user = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.471528] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.471700] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.471862] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.rescue_image_id = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.472045] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.rescue_kernel_id = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.472229] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.rescue_ramdisk_id = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.472404] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.472562] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.rx_queue_size = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.472728] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.smbfs_mount_options = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.473008] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.473190] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.snapshot_compression = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.473357] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.snapshot_image_format = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.473576] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.473742] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.sparse_logical_volumes = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.473906] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.swtpm_enabled = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.474086] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.swtpm_group = tss {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.474286] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.swtpm_user = tss {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.474460] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.sysinfo_serial = unique {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.474618] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.tb_cache_size = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.474774] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.tx_queue_size = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.474938] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.uid_maps = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.475116] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.use_virtio_for_bridges = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.475305] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.virt_type = kvm {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.475476] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.volume_clear = zero {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.475640] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.volume_clear_size = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.475807] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.volume_use_multipath = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.475965] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.vzstorage_cache_path = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.476150] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.476320] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.476485] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.476652] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.476921] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.477111] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.vzstorage_mount_user = stack {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.477283] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.477456] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.auth_section = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.477628] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.auth_type = password {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.477789] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.cafile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.477947] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.certfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.478130] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.collect_timing = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.478292] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.connect_retries = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.478449] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.connect_retry_delay = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.478617] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.default_floating_pool = public {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.478776] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.endpoint_override = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.478937] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.extension_sync_interval = 600 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.479108] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.http_retries = 3 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.479272] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.insecure = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.479430] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.keyfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.479587] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.max_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.479754] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.479912] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.min_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.480090] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.ovs_bridge = br-int {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.480256] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.physnets = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.480424] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.region_name = RegionOne {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.480581] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.retriable_status_codes = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.480751] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.service_metadata_proxy = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.480911] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.service_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.481092] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.service_type = network {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.481258] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.split_loggers = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.481418] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.status_code_retries = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.481573] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.status_code_retry_delay = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.481729] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.timeout = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.481904] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.482103] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] neutron.version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.482297] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] notifications.bdms_in_notifications = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.482479] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] notifications.default_level = INFO {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.482654] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] notifications.notification_format = unversioned {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.482819] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] notifications.notify_on_state_change = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.482994] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.483188] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] pci.alias = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.483361] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] pci.device_spec = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.483525] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] pci.report_in_placement = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.483697] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.auth_section = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.483869] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.auth_type = password {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.484044] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.484236] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.cafile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.484406] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.certfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.484569] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.collect_timing = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.484726] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.connect_retries = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.484882] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.connect_retry_delay = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.485061] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.default_domain_id = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.485231] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.default_domain_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.485396] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.domain_id = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.485551] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.domain_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.485708] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.endpoint_override = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.485866] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.insecure = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.486032] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.keyfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.486195] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.max_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.486352] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.min_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.486517] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.password = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.486674] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.project_domain_id = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.486837] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.project_domain_name = Default {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.487010] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.project_id = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.487189] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.project_name = service {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.487359] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.region_name = RegionOne {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.487520] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.retriable_status_codes = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.487680] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.service_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.487845] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.service_type = placement {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.488014] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.split_loggers = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.488183] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.status_code_retries = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.488349] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.status_code_retry_delay = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.488510] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.system_scope = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.488667] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.timeout = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.488825] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.trust_id = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.488981] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.user_domain_id = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.489162] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.user_domain_name = Default {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.489322] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.user_id = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.489491] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.username = nova {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.489669] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.489828] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] placement.version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.490015] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] quota.cores = 20 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.490188] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] quota.count_usage_from_placement = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.490364] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.490541] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] quota.injected_file_content_bytes = 10240 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.490710] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] quota.injected_file_path_length = 255 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.490879] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] quota.injected_files = 5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.491058] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] quota.instances = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.491233] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] quota.key_pairs = 100 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.491406] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] quota.metadata_items = 128 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.491572] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] quota.ram = 51200 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.491735] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] quota.recheck_quota = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.491904] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] quota.server_group_members = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.492100] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] quota.server_groups = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.492291] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.492461] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.492626] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] scheduler.image_metadata_prefilter = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.492789] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.492952] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] scheduler.max_attempts = 3 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.493130] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] scheduler.max_placement_results = 1000 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.493300] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.493461] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.493621] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.493795] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] scheduler.workers = 2 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.493966] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.494174] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.494371] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.494542] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.494707] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.494872] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.495048] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.495277] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.495456] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.host_subset_size = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.495621] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.495782] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.495945] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.496124] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.isolated_hosts = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.496291] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.isolated_images = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.496456] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.496614] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.496776] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.496935] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.pci_in_placement = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.497110] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.497279] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.497441] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.497601] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.497761] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.497921] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.498093] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.track_instance_changes = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.498276] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.498448] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] metrics.required = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.498611] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] metrics.weight_multiplier = 1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.498780] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.498947] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] metrics.weight_setting = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.499277] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.499457] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] serial_console.enabled = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.499635] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] serial_console.port_range = 10000:20000 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.499807] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.499978] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.500165] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] serial_console.serialproxy_port = 6083 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.500337] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] service_user.auth_section = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.500509] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] service_user.auth_type = password {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.500670] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] service_user.cafile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.500827] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] service_user.certfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.500988] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] service_user.collect_timing = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.501163] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] service_user.insecure = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.501324] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] service_user.keyfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.501492] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] service_user.send_service_user_token = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.501650] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] service_user.split_loggers = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.501824] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] service_user.timeout = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.501993] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] spice.agent_enabled = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.502198] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] spice.enabled = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.502514] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.502703] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.502873] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] spice.html5proxy_port = 6082 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.503046] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] spice.image_compression = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.503210] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] spice.jpeg_compression = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.503369] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] spice.playback_compression = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.503530] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] spice.require_secure = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.503699] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] spice.server_listen = 127.0.0.1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.503866] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.504035] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] spice.streaming_mode = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.504219] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] spice.zlib_compression = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.504396] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] upgrade_levels.baseapi = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.504569] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] upgrade_levels.compute = auto {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.504729] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] upgrade_levels.conductor = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.504885] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] upgrade_levels.scheduler = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.505062] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.505233] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.505391] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vendordata_dynamic_auth.cafile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.505546] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vendordata_dynamic_auth.certfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.505704] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.505862] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vendordata_dynamic_auth.insecure = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.506028] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.506195] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.506354] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vendordata_dynamic_auth.timeout = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.506529] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.api_retry_count = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.506688] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.ca_file = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.506858] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.507035] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.cluster_name = testcl1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.507202] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.connection_pool_size = 10 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.507362] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.console_delay_seconds = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.507532] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.datastore_regex = ^datastore.* {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.507735] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.507906] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.host_password = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.508082] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.host_port = 443 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.508255] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.host_username = administrator@vsphere.local {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.508425] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.insecure = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.508585] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.integration_bridge = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.508748] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.maximum_objects = 100 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.508904] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.pbm_default_policy = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.509076] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.pbm_enabled = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.509238] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.pbm_wsdl_location = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.509408] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.509566] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.serial_port_proxy_uri = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.509722] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.serial_port_service_uri = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.509887] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.task_poll_interval = 0.5 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.510068] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.use_linked_clone = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.510243] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.vnc_keymap = en-us {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.510410] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.vnc_port = 5900 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.510572] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vmware.vnc_port_total = 10000 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.510756] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vnc.auth_schemes = ['none'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.510931] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vnc.enabled = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.511237] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.511429] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.511603] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vnc.novncproxy_port = 6080 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.511781] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vnc.server_listen = 127.0.0.1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.511951] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.512156] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vnc.vencrypt_ca_certs = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.512330] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vnc.vencrypt_client_cert = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.512493] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vnc.vencrypt_client_key = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.512674] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.512838] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.disable_deep_image_inspection = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.512999] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.513178] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.513344] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.513503] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.disable_rootwrap = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.513664] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.enable_numa_live_migration = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.513824] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.513986] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.514193] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.514366] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.libvirt_disable_apic = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.514529] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.514693] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.514853] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.515024] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.515192] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.515353] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.515515] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.515672] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.515830] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.515992] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.516192] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.516364] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] wsgi.client_socket_timeout = 900 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.516532] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] wsgi.default_pool_size = 1000 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.516695] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] wsgi.keep_alive = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.516864] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] wsgi.max_header_line = 16384 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.517037] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.517206] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] wsgi.ssl_ca_file = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.517368] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] wsgi.ssl_cert_file = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.517526] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] wsgi.ssl_key_file = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.517691] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] wsgi.tcp_keepidle = 600 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.517871] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.518050] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] zvm.ca_file = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.518217] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] zvm.cloud_connector_url = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.518505] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.518682] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] zvm.reachable_timeout = 300 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.518864] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_policy.enforce_new_defaults = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.519268] env[62914]: WARNING oslo_config.cfg [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 471.519459] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_policy.enforce_scope = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.519639] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_policy.policy_default_rule = default {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.519823] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.520006] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_policy.policy_file = policy.yaml {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.520191] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.520359] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.520520] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.520677] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.520838] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.521020] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.521197] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.521376] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] profiler.connection_string = messaging:// {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.521544] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] profiler.enabled = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.521985] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] profiler.es_doc_type = notification {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.521985] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] profiler.es_scroll_size = 10000 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.522062] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] profiler.es_scroll_time = 2m {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.522244] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] profiler.filter_error_trace = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.522420] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] profiler.hmac_keys = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.522591] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] profiler.sentinel_service_name = mymaster {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.522758] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] profiler.socket_timeout = 0.1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.522921] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] profiler.trace_requests = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.523095] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] profiler.trace_sqlalchemy = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.523283] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] profiler_jaeger.process_tags = {} {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.523450] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] profiler_jaeger.service_name_prefix = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.523615] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] profiler_otlp.service_name_prefix = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.523780] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] remote_debug.host = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.523938] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] remote_debug.port = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.524162] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.524337] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.524502] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.524663] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.524828] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.524985] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.525161] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.525321] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.525483] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.525658] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.525815] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.525988] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.526171] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.526345] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.526519] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.526684] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.526846] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.527130] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.527201] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.527354] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.527530] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.527694] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.527856] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.528030] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.528198] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.528361] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.528521] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.528680] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.528845] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.529017] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.ssl = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.529196] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.529371] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.529533] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.529706] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.529875] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.530049] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.530252] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.530414] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_notifications.retry = -1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.530597] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.530773] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.530946] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.auth_section = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.531122] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.auth_type = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.531287] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.cafile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.531446] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.certfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.531608] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.collect_timing = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.531766] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.connect_retries = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.531923] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.connect_retry_delay = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.532235] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.endpoint_id = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.532281] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.endpoint_override = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.532444] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.insecure = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.532599] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.keyfile = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.532752] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.max_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.532909] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.min_version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.533077] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.region_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.533240] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.retriable_status_codes = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.533403] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.service_name = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.533561] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.service_type = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.533725] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.split_loggers = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.533885] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.status_code_retries = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.534053] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.status_code_retry_delay = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.534246] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.timeout = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.534416] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.valid_interfaces = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.534573] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_limit.version = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.534739] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_reports.file_event_handler = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.534903] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.535073] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] oslo_reports.log_dir = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.535250] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.535414] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.535573] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.535741] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.535907] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.536078] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.536250] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.536409] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vif_plug_ovs_privileged.group = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.536564] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.536727] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.536885] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.537051] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] vif_plug_ovs_privileged.user = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.537225] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.537404] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.537575] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.537743] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.537912] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.538089] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.538258] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.538422] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.538602] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.538774] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_ovs.isolate_vif = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.538938] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.539115] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.539287] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.539456] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.539617] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] os_vif_ovs.per_port_bridge = False {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.539789] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] privsep_osbrick.capabilities = [21] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.539948] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] privsep_osbrick.group = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.540121] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] privsep_osbrick.helper_command = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.540291] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.540452] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.540609] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] privsep_osbrick.user = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.540779] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.540935] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] nova_sys_admin.group = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.541105] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] nova_sys_admin.helper_command = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.541272] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.541434] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.541590] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] nova_sys_admin.user = None {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.541719] env[62914]: DEBUG oslo_service.service [None req-28dbd7fa-e245-4c9e-810b-05fa5fff55c0 None None] ******************************************************************************** {{(pid=62914) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 471.542258] env[62914]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 472.045643] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Getting list of instances from cluster (obj){ [ 472.045643] env[62914]: value = "domain-c8" [ 472.045643] env[62914]: _type = "ClusterComputeResource" [ 472.045643] env[62914]: } {{(pid=62914) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 472.046833] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900c5765-4a83-4a9a-9e3a-ead344087153 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.055932] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Got total of 0 instances {{(pid=62914) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 472.056644] env[62914]: WARNING nova.virt.vmwareapi.driver [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 472.057113] env[62914]: INFO nova.virt.node [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Generated node identity 8e1b3185-5ae3-4780-aee8-c507b157ad13 [ 472.057357] env[62914]: INFO nova.virt.node [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Wrote node identity 8e1b3185-5ae3-4780-aee8-c507b157ad13 to /opt/stack/data/n-cpu-1/compute_id [ 472.561039] env[62914]: WARNING nova.compute.manager [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Compute nodes ['8e1b3185-5ae3-4780-aee8-c507b157ad13'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 473.566117] env[62914]: INFO nova.compute.manager [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 474.570931] env[62914]: WARNING nova.compute.manager [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 474.571305] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 474.571449] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 474.571565] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 474.572032] env[62914]: DEBUG nova.compute.resource_tracker [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62914) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 474.572694] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384de480-2b19-4321-aecf-eaff20454c3f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.581204] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de0824c-09c3-4c34-8621-8de46c0b48a2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.594868] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9ab206-9858-4ec0-85c5-1fbd47bd8617 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.601393] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f0f22e8-2540-47b5-acd8-83c827e0c073 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.630803] env[62914]: DEBUG nova.compute.resource_tracker [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181259MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=62914) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 474.630971] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 474.631125] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 475.133698] env[62914]: WARNING nova.compute.resource_tracker [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] No compute node record for cpu-1:8e1b3185-5ae3-4780-aee8-c507b157ad13: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 8e1b3185-5ae3-4780-aee8-c507b157ad13 could not be found. [ 475.637624] env[62914]: INFO nova.compute.resource_tracker [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 8e1b3185-5ae3-4780-aee8-c507b157ad13 [ 477.146128] env[62914]: DEBUG nova.compute.resource_tracker [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 477.146485] env[62914]: DEBUG nova.compute.resource_tracker [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 477.297527] env[62914]: INFO nova.scheduler.client.report [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] [req-d9992085-f606-4f5c-a840-f32741609cda] Created resource provider record via placement API for resource provider with UUID 8e1b3185-5ae3-4780-aee8-c507b157ad13 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 477.313514] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be7132af-dfbc-48df-acdc-b94424aa54d8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.321293] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08547f9-585d-4ae9-b5e6-b22c55e7ddc7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.352724] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f1b539c-e003-4184-b132-d57dd1abea5a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.360153] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d532534-2a94-43e3-af1d-e6416672a8ef {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.373252] env[62914]: DEBUG nova.compute.provider_tree [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Updating inventory in ProviderTree for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 477.908580] env[62914]: DEBUG nova.scheduler.client.report [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Updated inventory for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 477.908815] env[62914]: DEBUG nova.compute.provider_tree [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Updating resource provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 generation from 0 to 1 during operation: update_inventory {{(pid=62914) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 477.908953] env[62914]: DEBUG nova.compute.provider_tree [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Updating inventory in ProviderTree for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 477.959174] env[62914]: DEBUG nova.compute.provider_tree [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Updating resource provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 generation from 1 to 2 during operation: update_traits {{(pid=62914) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 478.463869] env[62914]: DEBUG nova.compute.resource_tracker [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 478.464302] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.833s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 478.464378] env[62914]: DEBUG nova.service [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Creating RPC server for service compute {{(pid=62914) start /opt/stack/nova/nova/service.py:186}} [ 478.479866] env[62914]: DEBUG nova.service [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] Join ServiceGroup membership for this service compute {{(pid=62914) start /opt/stack/nova/nova/service.py:203}} [ 478.480058] env[62914]: DEBUG nova.servicegroup.drivers.db [None req-f69e1074-1358-49ce-93be-90c57ce4e9d7 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62914) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 491.483872] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._sync_power_states {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 491.987525] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Getting list of instances from cluster (obj){ [ 491.987525] env[62914]: value = "domain-c8" [ 491.987525] env[62914]: _type = "ClusterComputeResource" [ 491.987525] env[62914]: } {{(pid=62914) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 491.988763] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d07a545-4666-43a4-ab33-db94fa451f4a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 491.997481] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Got total of 0 instances {{(pid=62914) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 491.997728] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 491.998056] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Getting list of instances from cluster (obj){ [ 491.998056] env[62914]: value = "domain-c8" [ 491.998056] env[62914]: _type = "ClusterComputeResource" [ 491.998056] env[62914]: } {{(pid=62914) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 491.998935] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b7987d-49ab-4bf7-8056-94520854ba40 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.005939] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Got total of 0 instances {{(pid=62914) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 513.359906] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Acquiring lock "b100dbb2-7e2e-43da-bb5f-e31c6676d1bd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.361633] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Lock "b100dbb2-7e2e-43da-bb5f-e31c6676d1bd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.864310] env[62914]: DEBUG nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 514.435389] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.436110] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.438292] env[62914]: INFO nova.compute.claims [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 515.320817] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Acquiring lock "3db8507b-52b8-4eeb-b801-56e817c3b728" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.321108] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Lock "3db8507b-52b8-4eeb-b801-56e817c3b728" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.517479] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ef1488-bc0d-4545-8331-381e99abd229 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.530020] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1233fa6-be4b-4ccb-903b-016aa393eb89 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.579549] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156f49df-a044-4d6b-a554-bac5008364dd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.588861] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9765a202-a3f8-483b-88fe-f75f5b9e71a5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.606122] env[62914]: DEBUG nova.compute.provider_tree [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 515.805258] env[62914]: DEBUG oslo_concurrency.lockutils [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquiring lock "3786cd93-1a50-4997-8894-840afa1c8417" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.805533] env[62914]: DEBUG oslo_concurrency.lockutils [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Lock "3786cd93-1a50-4997-8894-840afa1c8417" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.823468] env[62914]: DEBUG nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 516.112058] env[62914]: DEBUG nova.scheduler.client.report [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 516.312491] env[62914]: DEBUG nova.compute.manager [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 516.358386] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.619252] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.183s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 516.619961] env[62914]: DEBUG nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 516.626313] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.268s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.628073] env[62914]: INFO nova.compute.claims [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 516.777125] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Acquiring lock "33f4159e-4181-4345-a716-0dfc409cee78" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.777342] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Lock "33f4159e-4181-4345-a716-0dfc409cee78" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.840496] env[62914]: DEBUG oslo_concurrency.lockutils [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.132542] env[62914]: DEBUG nova.compute.utils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 517.134486] env[62914]: DEBUG nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 517.134748] env[62914]: DEBUG nova.network.neutron [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 517.281762] env[62914]: DEBUG nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.421956] env[62914]: DEBUG nova.policy [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7151f7b1aa7b4be69a0fda545fa32957', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '885472ede2ff409c9a2f1e434c812fd3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 517.645023] env[62914]: DEBUG nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 517.780697] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-400879eb-50a0-438d-a620-8d3295735edd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.794210] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933f0839-b0a8-4a8d-92b6-de83a0316956 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.832927] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Acquiring lock "d5c90d13-41bc-4cca-abf3-5a5b13171a6f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.833240] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Lock "d5c90d13-41bc-4cca-abf3-5a5b13171a6f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.834366] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Acquiring lock "a7ee1dbd-73d2-46be-bf43-008eebcd64e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.834589] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Lock "a7ee1dbd-73d2-46be-bf43-008eebcd64e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.837803] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.838949] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd96dcc-995b-4175-8e2e-b3dc282e085d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.847282] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab17065c-3338-4c09-8d33-ac9b47424132 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.864178] env[62914]: DEBUG nova.compute.provider_tree [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 518.339277] env[62914]: DEBUG nova.compute.manager [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.355182] env[62914]: DEBUG nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.367107] env[62914]: DEBUG nova.scheduler.client.report [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 518.656551] env[62914]: DEBUG nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 518.696711] env[62914]: DEBUG nova.virt.hardware [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 518.697089] env[62914]: DEBUG nova.virt.hardware [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 518.697089] env[62914]: DEBUG nova.virt.hardware [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 518.697538] env[62914]: DEBUG nova.virt.hardware [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 518.699954] env[62914]: DEBUG nova.virt.hardware [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 518.699954] env[62914]: DEBUG nova.virt.hardware [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 518.700568] env[62914]: DEBUG nova.virt.hardware [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 518.700568] env[62914]: DEBUG nova.virt.hardware [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 518.701148] env[62914]: DEBUG nova.virt.hardware [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 518.701148] env[62914]: DEBUG nova.virt.hardware [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 518.701369] env[62914]: DEBUG nova.virt.hardware [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 518.702420] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f394fa1-38d8-405a-8c98-cd9bc0ceb4d5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.714545] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97af5bb8-e7b7-4c03-b5f9-2b22e6ffc9a6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.734558] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7b8a5bd-1015-4a27-8cda-75cf7ea8a0f3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.764195] env[62914]: DEBUG nova.network.neutron [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Successfully created port: 6fde6d50-bd0f-4768-911b-5b20080bdfa3 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 518.876548] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.878451] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.252s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 518.880189] env[62914]: DEBUG nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 518.886505] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.886812] env[62914]: DEBUG oslo_concurrency.lockutils [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.046s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.888295] env[62914]: INFO nova.compute.claims [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 519.397931] env[62914]: DEBUG nova.compute.utils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 519.400684] env[62914]: DEBUG nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 519.400756] env[62914]: DEBUG nova.network.neutron [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 519.831547] env[62914]: DEBUG nova.policy [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93ae51875f3949cc8be30c571a55e36a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8661754805184e2abf74c6fb8b54fe5a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 519.907725] env[62914]: DEBUG nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 520.050393] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c74a18-6d95-44ed-83ef-11c75641dfe3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.060557] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d79512f-069d-41b5-9fdb-25f3838df156 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.096225] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58006fb-e48e-40df-9c01-f6b6702ea0f9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.104442] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6031139-e536-4c7b-94f5-d4c282e60b87 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.122539] env[62914]: DEBUG nova.compute.provider_tree [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 520.634108] env[62914]: DEBUG nova.scheduler.client.report [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 520.923644] env[62914]: DEBUG nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 520.966018] env[62914]: DEBUG nova.virt.hardware [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 520.966018] env[62914]: DEBUG nova.virt.hardware [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 520.966018] env[62914]: DEBUG nova.virt.hardware [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 520.966348] env[62914]: DEBUG nova.virt.hardware [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 520.966348] env[62914]: DEBUG nova.virt.hardware [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 520.966348] env[62914]: DEBUG nova.virt.hardware [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 520.966348] env[62914]: DEBUG nova.virt.hardware [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 520.966348] env[62914]: DEBUG nova.virt.hardware [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 520.966514] env[62914]: DEBUG nova.virt.hardware [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 520.966514] env[62914]: DEBUG nova.virt.hardware [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 520.966514] env[62914]: DEBUG nova.virt.hardware [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 520.966868] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27f5c76-4843-4069-ad87-678ef204811b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.975740] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f47c58-86cc-4554-ad8c-972112346aff {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.995871] env[62914]: DEBUG nova.network.neutron [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Successfully created port: ade1c8df-fe65-4c76-9615-8490472a1ed7 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 521.140155] env[62914]: DEBUG oslo_concurrency.lockutils [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 521.141613] env[62914]: DEBUG nova.compute.manager [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 521.146147] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.307s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.146147] env[62914]: INFO nova.compute.claims [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 521.653139] env[62914]: DEBUG nova.compute.utils [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 521.658255] env[62914]: DEBUG nova.compute.manager [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Not allocating networking since 'none' was specified. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 522.160680] env[62914]: DEBUG nova.compute.manager [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 522.284581] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2aa940d-975a-4261-a27d-e3098ad39040 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.292854] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b8fcc7-ab80-4f2b-b6c3-808f46f6eb56 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.331795] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e5efb34-66d3-43c0-92ed-ca9779d7b7e8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.339569] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc82ddf5-99e7-49ef-9dd5-efa4b93fca70 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.357924] env[62914]: DEBUG nova.compute.provider_tree [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 522.863572] env[62914]: DEBUG nova.scheduler.client.report [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 523.024525] env[62914]: ERROR nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6fde6d50-bd0f-4768-911b-5b20080bdfa3, please check neutron logs for more information. [ 523.024525] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 523.024525] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.024525] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 523.024525] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.024525] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 523.024525] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.024525] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 523.024525] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.024525] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 523.024525] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.024525] env[62914]: ERROR nova.compute.manager raise self.value [ 523.024525] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.024525] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 523.024525] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.024525] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 523.025231] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.025231] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 523.025231] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6fde6d50-bd0f-4768-911b-5b20080bdfa3, please check neutron logs for more information. [ 523.025231] env[62914]: ERROR nova.compute.manager [ 523.025490] env[62914]: Traceback (most recent call last): [ 523.025725] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 523.025725] env[62914]: listener.cb(fileno) [ 523.025725] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.025725] env[62914]: result = function(*args, **kwargs) [ 523.025725] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 523.025725] env[62914]: return func(*args, **kwargs) [ 523.025725] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.025725] env[62914]: raise e [ 523.025725] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.025725] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 523.025725] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.025725] env[62914]: created_port_ids = self._update_ports_for_instance( [ 523.025725] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.025725] env[62914]: with excutils.save_and_reraise_exception(): [ 523.025725] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.025725] env[62914]: self.force_reraise() [ 523.025725] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.025725] env[62914]: raise self.value [ 523.025725] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.025725] env[62914]: updated_port = self._update_port( [ 523.025725] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.025725] env[62914]: _ensure_no_port_binding_failure(port) [ 523.025725] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.025725] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 523.025725] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 6fde6d50-bd0f-4768-911b-5b20080bdfa3, please check neutron logs for more information. [ 523.025725] env[62914]: Removing descriptor: 14 [ 523.028847] env[62914]: ERROR nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6fde6d50-bd0f-4768-911b-5b20080bdfa3, please check neutron logs for more information. [ 523.028847] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Traceback (most recent call last): [ 523.028847] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 523.028847] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] yield resources [ 523.028847] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 523.028847] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] self.driver.spawn(context, instance, image_meta, [ 523.028847] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 523.028847] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.028847] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.028847] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] vm_ref = self.build_virtual_machine(instance, [ 523.028847] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.029240] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.029240] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.029240] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] for vif in network_info: [ 523.029240] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 523.029240] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] return self._sync_wrapper(fn, *args, **kwargs) [ 523.029240] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 523.029240] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] self.wait() [ 523.029240] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 523.029240] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] self[:] = self._gt.wait() [ 523.029240] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.029240] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] return self._exit_event.wait() [ 523.029240] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.029240] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] result = hub.switch() [ 523.030256] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.030256] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] return self.greenlet.switch() [ 523.030256] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.030256] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] result = function(*args, **kwargs) [ 523.030256] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 523.030256] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] return func(*args, **kwargs) [ 523.030256] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.030256] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] raise e [ 523.030256] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.030256] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] nwinfo = self.network_api.allocate_for_instance( [ 523.030256] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.030256] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] created_port_ids = self._update_ports_for_instance( [ 523.030256] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.030679] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] with excutils.save_and_reraise_exception(): [ 523.030679] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.030679] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] self.force_reraise() [ 523.030679] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.030679] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] raise self.value [ 523.030679] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.030679] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] updated_port = self._update_port( [ 523.030679] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.030679] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] _ensure_no_port_binding_failure(port) [ 523.030679] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.030679] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] raise exception.PortBindingFailed(port_id=port['id']) [ 523.030679] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] nova.exception.PortBindingFailed: Binding failed for port 6fde6d50-bd0f-4768-911b-5b20080bdfa3, please check neutron logs for more information. [ 523.030679] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] [ 523.031018] env[62914]: INFO nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Terminating instance [ 523.035484] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Acquiring lock "refresh_cache-b100dbb2-7e2e-43da-bb5f-e31c6676d1bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.035795] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Acquired lock "refresh_cache-b100dbb2-7e2e-43da-bb5f-e31c6676d1bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.036158] env[62914]: DEBUG nova.network.neutron [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 523.181886] env[62914]: DEBUG nova.compute.manager [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 523.211124] env[62914]: DEBUG nova.virt.hardware [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 523.211124] env[62914]: DEBUG nova.virt.hardware [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 523.211124] env[62914]: DEBUG nova.virt.hardware [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 523.211124] env[62914]: DEBUG nova.virt.hardware [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 523.211401] env[62914]: DEBUG nova.virt.hardware [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 523.211401] env[62914]: DEBUG nova.virt.hardware [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 523.211401] env[62914]: DEBUG nova.virt.hardware [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 523.211401] env[62914]: DEBUG nova.virt.hardware [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 523.215471] env[62914]: DEBUG nova.virt.hardware [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 523.215768] env[62914]: DEBUG nova.virt.hardware [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 523.215896] env[62914]: DEBUG nova.virt.hardware [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 523.216827] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8316b519-85ad-4e1c-8690-0b0c4a146cd1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.228316] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb94078e-9e51-4655-9413-0b05b0b38cce {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.245068] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Instance VIF info [] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 523.254884] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 523.254884] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-13d7b8c8-fca4-4a4f-bac2-816eeb2b5ce3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.268544] env[62914]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 523.268958] env[62914]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62914) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 523.269235] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 523.269235] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Creating folder: Project (ea30269c980e4477874063373e366dd9). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 523.269708] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7274df6f-c0df-4fd2-acae-c41d78670251 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.280358] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Created folder: Project (ea30269c980e4477874063373e366dd9) in parent group-v288131. [ 523.280576] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Creating folder: Instances. Parent ref: group-v288135. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 523.280826] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-17d77db8-2884-4d02-8edf-781b558b6045 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.292462] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Created folder: Instances in parent group-v288135. [ 523.292462] env[62914]: DEBUG oslo.service.loopingcall [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 523.292462] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 523.292462] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac5e0d28-5227-4cc9-a8d4-5a1b68bcee18 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.313981] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 523.313981] env[62914]: value = "task-1352311" [ 523.313981] env[62914]: _type = "Task" [ 523.313981] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.324493] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352311, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.335896] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Acquiring lock "c78502a7-266d-49cf-9d50-cc89ed266647" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.336053] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Lock "c78502a7-266d-49cf-9d50-cc89ed266647" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.372021] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.225s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.372021] env[62914]: DEBUG nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 523.375805] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.499s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.378128] env[62914]: INFO nova.compute.claims [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 523.600212] env[62914]: DEBUG nova.network.neutron [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.791128] env[62914]: DEBUG nova.network.neutron [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.826954] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352311, 'name': CreateVM_Task, 'duration_secs': 0.291618} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.827146] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 523.828200] env[62914]: DEBUG oslo_vmware.service [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db2a85e-a684-4b36-8d3a-f0d6d819a845 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.833751] env[62914]: ERROR nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ade1c8df-fe65-4c76-9615-8490472a1ed7, please check neutron logs for more information. [ 523.833751] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 523.833751] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.833751] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 523.833751] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.833751] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 523.833751] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.833751] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 523.833751] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.833751] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 523.833751] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.833751] env[62914]: ERROR nova.compute.manager raise self.value [ 523.833751] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.833751] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 523.833751] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.833751] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 523.834540] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.834540] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 523.834540] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ade1c8df-fe65-4c76-9615-8490472a1ed7, please check neutron logs for more information. [ 523.834540] env[62914]: ERROR nova.compute.manager [ 523.834540] env[62914]: Traceback (most recent call last): [ 523.834540] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 523.834540] env[62914]: listener.cb(fileno) [ 523.834540] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.834540] env[62914]: result = function(*args, **kwargs) [ 523.834540] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 523.834540] env[62914]: return func(*args, **kwargs) [ 523.834540] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.834540] env[62914]: raise e [ 523.834540] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.834540] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 523.834540] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.834540] env[62914]: created_port_ids = self._update_ports_for_instance( [ 523.834540] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.834540] env[62914]: with excutils.save_and_reraise_exception(): [ 523.834540] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.834540] env[62914]: self.force_reraise() [ 523.834540] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.834540] env[62914]: raise self.value [ 523.834540] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.834540] env[62914]: updated_port = self._update_port( [ 523.834540] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.834540] env[62914]: _ensure_no_port_binding_failure(port) [ 523.834540] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.834540] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 523.836585] env[62914]: nova.exception.PortBindingFailed: Binding failed for port ade1c8df-fe65-4c76-9615-8490472a1ed7, please check neutron logs for more information. [ 523.836585] env[62914]: Removing descriptor: 15 [ 523.836585] env[62914]: DEBUG oslo_concurrency.lockutils [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.836585] env[62914]: DEBUG oslo_concurrency.lockutils [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.836729] env[62914]: DEBUG oslo_concurrency.lockutils [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 523.837503] env[62914]: ERROR nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ade1c8df-fe65-4c76-9615-8490472a1ed7, please check neutron logs for more information. [ 523.837503] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Traceback (most recent call last): [ 523.837503] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 523.837503] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] yield resources [ 523.837503] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 523.837503] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] self.driver.spawn(context, instance, image_meta, [ 523.837503] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 523.837503] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.837503] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.837503] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] vm_ref = self.build_virtual_machine(instance, [ 523.837503] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.837816] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.837816] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.837816] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] for vif in network_info: [ 523.837816] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 523.837816] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] return self._sync_wrapper(fn, *args, **kwargs) [ 523.837816] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 523.837816] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] self.wait() [ 523.837816] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 523.837816] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] self[:] = self._gt.wait() [ 523.837816] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.837816] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] return self._exit_event.wait() [ 523.837816] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.837816] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] result = hub.switch() [ 523.838169] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.838169] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] return self.greenlet.switch() [ 523.838169] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.838169] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] result = function(*args, **kwargs) [ 523.838169] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 523.838169] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] return func(*args, **kwargs) [ 523.838169] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.838169] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] raise e [ 523.838169] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.838169] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] nwinfo = self.network_api.allocate_for_instance( [ 523.838169] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.838169] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] created_port_ids = self._update_ports_for_instance( [ 523.838169] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.838480] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] with excutils.save_and_reraise_exception(): [ 523.838480] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.838480] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] self.force_reraise() [ 523.838480] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.838480] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] raise self.value [ 523.838480] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.838480] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] updated_port = self._update_port( [ 523.838480] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.838480] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] _ensure_no_port_binding_failure(port) [ 523.838480] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.838480] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] raise exception.PortBindingFailed(port_id=port['id']) [ 523.838480] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] nova.exception.PortBindingFailed: Binding failed for port ade1c8df-fe65-4c76-9615-8490472a1ed7, please check neutron logs for more information. [ 523.838480] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] [ 523.838803] env[62914]: INFO nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Terminating instance [ 523.839947] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54591c94-f732-4f58-ad59-98e5eadfbdf7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.841960] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Acquiring lock "refresh_cache-3db8507b-52b8-4eeb-b801-56e817c3b728" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.842151] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Acquired lock "refresh_cache-3db8507b-52b8-4eeb-b801-56e817c3b728" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.842314] env[62914]: DEBUG nova.network.neutron [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 523.844448] env[62914]: DEBUG nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 523.849907] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 523.849907] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52336f66-605f-8aac-1699-fe349fa905a3" [ 523.849907] env[62914]: _type = "Task" [ 523.849907] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.858863] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52336f66-605f-8aac-1699-fe349fa905a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.883440] env[62914]: DEBUG nova.compute.utils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 523.885743] env[62914]: DEBUG nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 523.887022] env[62914]: DEBUG nova.network.neutron [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 523.979845] env[62914]: DEBUG nova.policy [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4f9eed574e974747b661c414b72c32d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '20575a9f5ae24c9dbe1127a84973efa4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 524.229710] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "df6047b8-8e4b-48e2-ae54-f5fda0aa77da" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.229967] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "df6047b8-8e4b-48e2-ae54-f5fda0aa77da" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.298479] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Releasing lock "refresh_cache-b100dbb2-7e2e-43da-bb5f-e31c6676d1bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.298927] env[62914]: DEBUG nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 524.299166] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 524.299797] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e327cc87-ceb1-4bd4-a4a4-b8ac0b95177a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.312275] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3abeca37-829a-401b-9d32-c23d17452bdb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.359321] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b100dbb2-7e2e-43da-bb5f-e31c6676d1bd could not be found. [ 524.359592] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 524.359842] env[62914]: INFO nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Took 0.06 seconds to destroy the instance on the hypervisor. [ 524.360158] env[62914]: DEBUG oslo.service.loopingcall [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 524.364705] env[62914]: DEBUG nova.compute.manager [-] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 524.364799] env[62914]: DEBUG nova.network.neutron [-] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 524.375712] env[62914]: DEBUG oslo_concurrency.lockutils [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.376030] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 524.378020] env[62914]: DEBUG oslo_concurrency.lockutils [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.378020] env[62914]: DEBUG oslo_concurrency.lockutils [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.378020] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 524.378020] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b4412f1-e0e5-4e2f-83cd-a1e7f9791923 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.386510] env[62914]: DEBUG nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 524.402906] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 524.403129] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 524.404044] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa35fd96-0464-4c76-8ca9-2749ae002a0e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.417817] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.427047] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0ea4c1c-12fd-4e5a-98bf-13282565c797 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.437446] env[62914]: DEBUG nova.network.neutron [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.437446] env[62914]: DEBUG nova.network.neutron [-] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.445682] env[62914]: DEBUG nova.compute.manager [req-fef47b6e-6fba-4243-b175-83a6c2817b6c req-745ea8c7-c0cb-4c18-a5ea-98dea3164fad service nova] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Received event network-changed-6fde6d50-bd0f-4768-911b-5b20080bdfa3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 524.445912] env[62914]: DEBUG nova.compute.manager [req-fef47b6e-6fba-4243-b175-83a6c2817b6c req-745ea8c7-c0cb-4c18-a5ea-98dea3164fad service nova] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Refreshing instance network info cache due to event network-changed-6fde6d50-bd0f-4768-911b-5b20080bdfa3. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 524.447702] env[62914]: DEBUG oslo_concurrency.lockutils [req-fef47b6e-6fba-4243-b175-83a6c2817b6c req-745ea8c7-c0cb-4c18-a5ea-98dea3164fad service nova] Acquiring lock "refresh_cache-b100dbb2-7e2e-43da-bb5f-e31c6676d1bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.447881] env[62914]: DEBUG oslo_concurrency.lockutils [req-fef47b6e-6fba-4243-b175-83a6c2817b6c req-745ea8c7-c0cb-4c18-a5ea-98dea3164fad service nova] Acquired lock "refresh_cache-b100dbb2-7e2e-43da-bb5f-e31c6676d1bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.448056] env[62914]: DEBUG nova.network.neutron [req-fef47b6e-6fba-4243-b175-83a6c2817b6c req-745ea8c7-c0cb-4c18-a5ea-98dea3164fad service nova] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Refreshing network info cache for port 6fde6d50-bd0f-4768-911b-5b20080bdfa3 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 524.458293] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 524.458293] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5202417b-1289-46cb-6bf6-9f41c0bed8c4" [ 524.458293] env[62914]: _type = "Task" [ 524.458293] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.473779] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Preparing fetch location {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 524.475215] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Creating directory with path [datastore2] vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116/6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 524.475685] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2e37a92f-a2a1-4fa2-a47c-80eac023757b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.503764] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Created directory with path [datastore2] vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116/6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 524.503974] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Fetch image to [datastore2] vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 524.504165] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Downloading image file data 6c21e11e-da68-48a7-9929-737212ce98cf to [datastore2] vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk on the data store datastore2 {{(pid=62914) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 524.505024] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d13887-67f9-4f53-a6e6-97ca08662fe3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.520747] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb4a027-2017-4c35-a0bf-ef7c78978972 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.539427] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34c23df0-6da2-4f8d-b280-6d74813cafe4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.585081] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca117590-f0d5-4da3-9aa8-8bfc41c22030 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.594138] env[62914]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-76717f2a-2469-477d-ab0c-32f0a6c9d6f5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.626881] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Acquiring lock "16b4e8bd-d84f-45a6-b531-3200ea545931" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.627189] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Lock "16b4e8bd-d84f-45a6-b531-3200ea545931" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.629078] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Downloading image file data 6c21e11e-da68-48a7-9929-737212ce98cf to the data store datastore2 {{(pid=62914) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 524.690825] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ae7d34-1605-4f06-8bc8-1b1b71afeec2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.704824] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eab01d5e-f75f-4940-a178-8ccbe009e887 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.743425] env[62914]: DEBUG nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 524.750021] env[62914]: DEBUG oslo_vmware.rw_handles [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62914) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 524.750021] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8b87d5-d703-4948-a25d-91e0c4e95a62 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.830015] env[62914]: DEBUG nova.network.neutron [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Successfully created port: 44ec2ba3-6f10-4fff-8e78-9872e8b8c0a0 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 524.838175] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c537d1-e062-43a2-870f-7905aa65bcb5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.856180] env[62914]: DEBUG nova.compute.provider_tree [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 524.942363] env[62914]: DEBUG nova.network.neutron [-] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.001743] env[62914]: DEBUG nova.network.neutron [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.021605] env[62914]: DEBUG nova.network.neutron [req-fef47b6e-6fba-4243-b175-83a6c2817b6c req-745ea8c7-c0cb-4c18-a5ea-98dea3164fad service nova] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 525.129996] env[62914]: DEBUG nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 525.225314] env[62914]: DEBUG nova.network.neutron [req-fef47b6e-6fba-4243-b175-83a6c2817b6c req-745ea8c7-c0cb-4c18-a5ea-98dea3164fad service nova] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.277343] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.359473] env[62914]: DEBUG nova.scheduler.client.report [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 525.404102] env[62914]: DEBUG nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 525.437995] env[62914]: DEBUG nova.virt.hardware [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 525.437995] env[62914]: DEBUG nova.virt.hardware [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 525.438151] env[62914]: DEBUG nova.virt.hardware [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 525.439799] env[62914]: DEBUG nova.virt.hardware [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 525.440101] env[62914]: DEBUG nova.virt.hardware [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 525.440248] env[62914]: DEBUG nova.virt.hardware [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 525.440512] env[62914]: DEBUG nova.virt.hardware [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 525.440746] env[62914]: DEBUG nova.virt.hardware [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 525.440828] env[62914]: DEBUG nova.virt.hardware [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 525.441049] env[62914]: DEBUG nova.virt.hardware [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 525.441329] env[62914]: DEBUG nova.virt.hardware [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 525.442182] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9baac5-5e59-40b2-abbc-5adbbe1ae0b8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.448406] env[62914]: INFO nova.compute.manager [-] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Took 1.08 seconds to deallocate network for instance. [ 525.454523] env[62914]: DEBUG nova.compute.claims [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 525.454523] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.459220] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd4d4974-311d-407f-91ae-74d492c22f02 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.505383] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Releasing lock "refresh_cache-3db8507b-52b8-4eeb-b801-56e817c3b728" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.505614] env[62914]: DEBUG nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 525.505834] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 525.506135] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2ebc393b-d0f7-442d-9844-b5e5a1b5ae19 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.526544] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e811caf9-8520-490d-9a9b-a63c275dbdc8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.549559] env[62914]: DEBUG oslo_vmware.rw_handles [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Completed reading data from the image iterator. {{(pid=62914) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 525.549559] env[62914]: DEBUG oslo_vmware.rw_handles [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62914) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 525.555376] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3db8507b-52b8-4eeb-b801-56e817c3b728 could not be found. [ 525.555493] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 525.555717] env[62914]: INFO nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Took 0.05 seconds to destroy the instance on the hypervisor. [ 525.558025] env[62914]: DEBUG oslo.service.loopingcall [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 525.558025] env[62914]: DEBUG nova.compute.manager [-] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 525.558025] env[62914]: DEBUG nova.network.neutron [-] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 525.600021] env[62914]: DEBUG nova.network.neutron [-] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 525.640461] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Downloaded image file data 6c21e11e-da68-48a7-9929-737212ce98cf to vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk on the data store datastore2 {{(pid=62914) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 525.642261] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Caching image {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 525.642535] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Copying Virtual Disk [datastore2] vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk to [datastore2] vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 525.644265] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b01fd48-7af2-4d20-bd46-48ce32716c9c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.657060] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 525.657060] env[62914]: value = "task-1352313" [ 525.657060] env[62914]: _type = "Task" [ 525.657060] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.664094] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352313, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.682485] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.731313] env[62914]: DEBUG oslo_concurrency.lockutils [req-fef47b6e-6fba-4243-b175-83a6c2817b6c req-745ea8c7-c0cb-4c18-a5ea-98dea3164fad service nova] Releasing lock "refresh_cache-b100dbb2-7e2e-43da-bb5f-e31c6676d1bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.865837] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.866102] env[62914]: DEBUG nova.compute.manager [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 525.868858] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.982s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.871386] env[62914]: INFO nova.compute.claims [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 526.106995] env[62914]: DEBUG nova.network.neutron [-] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.167878] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352313, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.381756] env[62914]: DEBUG nova.compute.utils [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 526.392159] env[62914]: DEBUG nova.compute.manager [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Not allocating networking since 'none' was specified. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 526.610677] env[62914]: INFO nova.compute.manager [-] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Took 1.05 seconds to deallocate network for instance. [ 526.617622] env[62914]: DEBUG nova.compute.claims [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 526.617622] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.668521] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352313, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.794203} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.668611] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Copied Virtual Disk [datastore2] vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk to [datastore2] vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 526.668835] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Deleting the datastore file [datastore2] vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 526.669441] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c251ecf2-9821-434d-b229-45cb17514daa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.676017] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 526.676017] env[62914]: value = "task-1352315" [ 526.676017] env[62914]: _type = "Task" [ 526.676017] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.684511] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352315, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.747018] env[62914]: ERROR nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 44ec2ba3-6f10-4fff-8e78-9872e8b8c0a0, please check neutron logs for more information. [ 526.747018] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 526.747018] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.747018] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 526.747018] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.747018] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 526.747018] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.747018] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 526.747018] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.747018] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 526.747018] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.747018] env[62914]: ERROR nova.compute.manager raise self.value [ 526.747018] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.747018] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 526.747018] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.747018] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 526.747791] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.747791] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 526.747791] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 44ec2ba3-6f10-4fff-8e78-9872e8b8c0a0, please check neutron logs for more information. [ 526.747791] env[62914]: ERROR nova.compute.manager [ 526.747791] env[62914]: Traceback (most recent call last): [ 526.747791] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 526.747791] env[62914]: listener.cb(fileno) [ 526.747791] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.747791] env[62914]: result = function(*args, **kwargs) [ 526.747791] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 526.747791] env[62914]: return func(*args, **kwargs) [ 526.747791] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 526.747791] env[62914]: raise e [ 526.747791] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.747791] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 526.747791] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.747791] env[62914]: created_port_ids = self._update_ports_for_instance( [ 526.747791] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.747791] env[62914]: with excutils.save_and_reraise_exception(): [ 526.747791] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.747791] env[62914]: self.force_reraise() [ 526.747791] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.747791] env[62914]: raise self.value [ 526.747791] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.747791] env[62914]: updated_port = self._update_port( [ 526.747791] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.747791] env[62914]: _ensure_no_port_binding_failure(port) [ 526.747791] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.747791] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 526.748852] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 44ec2ba3-6f10-4fff-8e78-9872e8b8c0a0, please check neutron logs for more information. [ 526.748852] env[62914]: Removing descriptor: 15 [ 526.748852] env[62914]: ERROR nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 44ec2ba3-6f10-4fff-8e78-9872e8b8c0a0, please check neutron logs for more information. [ 526.748852] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Traceback (most recent call last): [ 526.748852] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 526.748852] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] yield resources [ 526.748852] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 526.748852] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] self.driver.spawn(context, instance, image_meta, [ 526.748852] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 526.748852] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] self._vmops.spawn(context, instance, image_meta, injected_files, [ 526.748852] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 526.748852] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] vm_ref = self.build_virtual_machine(instance, [ 526.749182] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 526.749182] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] vif_infos = vmwarevif.get_vif_info(self._session, [ 526.749182] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 526.749182] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] for vif in network_info: [ 526.749182] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 526.749182] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] return self._sync_wrapper(fn, *args, **kwargs) [ 526.749182] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 526.749182] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] self.wait() [ 526.749182] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 526.749182] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] self[:] = self._gt.wait() [ 526.749182] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 526.749182] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] return self._exit_event.wait() [ 526.749182] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 526.749523] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] result = hub.switch() [ 526.749523] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 526.749523] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] return self.greenlet.switch() [ 526.749523] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.749523] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] result = function(*args, **kwargs) [ 526.749523] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 526.749523] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] return func(*args, **kwargs) [ 526.749523] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 526.749523] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] raise e [ 526.749523] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.749523] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] nwinfo = self.network_api.allocate_for_instance( [ 526.749523] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.749523] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] created_port_ids = self._update_ports_for_instance( [ 526.749913] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.749913] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] with excutils.save_and_reraise_exception(): [ 526.749913] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.749913] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] self.force_reraise() [ 526.749913] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.749913] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] raise self.value [ 526.749913] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.749913] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] updated_port = self._update_port( [ 526.749913] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.749913] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] _ensure_no_port_binding_failure(port) [ 526.749913] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.749913] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] raise exception.PortBindingFailed(port_id=port['id']) [ 526.750248] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] nova.exception.PortBindingFailed: Binding failed for port 44ec2ba3-6f10-4fff-8e78-9872e8b8c0a0, please check neutron logs for more information. [ 526.750248] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] [ 526.750248] env[62914]: INFO nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Terminating instance [ 526.751656] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Acquiring lock "refresh_cache-33f4159e-4181-4345-a716-0dfc409cee78" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.751971] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Acquired lock "refresh_cache-33f4159e-4181-4345-a716-0dfc409cee78" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.751971] env[62914]: DEBUG nova.network.neutron [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 526.898478] env[62914]: DEBUG nova.compute.manager [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 526.910610] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.910966] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.911231] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Starting heal instance info cache {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 526.911389] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Rebuilding the list of instances to heal {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 527.089301] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc89cf1-9be8-4a25-b4d9-f260ef6adcd8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.099996] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2c8dbd-7b8c-4b7c-b142-55a28c495366 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.134709] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d021117-aacd-4be8-aa7e-7b75bd0572cb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.144397] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee28076f-f3b5-4e8a-bbc5-28dc49ecad34 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.159084] env[62914]: DEBUG nova.compute.provider_tree [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Updating inventory in ProviderTree for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 527.187241] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352315, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026067} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.187540] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 527.187766] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Moving file from [datastore2] vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116/6c21e11e-da68-48a7-9929-737212ce98cf to [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf. {{(pid=62914) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 527.188043] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-6c1610bb-c91e-46ca-80d4-e7f06c1f221d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.195747] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 527.195747] env[62914]: value = "task-1352316" [ 527.195747] env[62914]: _type = "Task" [ 527.195747] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.207060] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352316, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.278631] env[62914]: DEBUG nova.network.neutron [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 527.365417] env[62914]: DEBUG nova.compute.manager [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Received event network-vif-deleted-6fde6d50-bd0f-4768-911b-5b20080bdfa3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.365916] env[62914]: DEBUG nova.compute.manager [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Received event network-changed-ade1c8df-fe65-4c76-9615-8490472a1ed7 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.366028] env[62914]: DEBUG nova.compute.manager [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Refreshing instance network info cache due to event network-changed-ade1c8df-fe65-4c76-9615-8490472a1ed7. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 527.366236] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] Acquiring lock "refresh_cache-3db8507b-52b8-4eeb-b801-56e817c3b728" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.366460] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] Acquired lock "refresh_cache-3db8507b-52b8-4eeb-b801-56e817c3b728" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.366532] env[62914]: DEBUG nova.network.neutron [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Refreshing network info cache for port ade1c8df-fe65-4c76-9615-8490472a1ed7 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 527.368360] env[62914]: DEBUG nova.network.neutron [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.416708] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 527.416708] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 527.416708] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 527.416708] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 527.416708] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 527.416708] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 527.417149] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Didn't find any instances for network info cache update. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 527.417149] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.417149] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.417501] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.418094] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.418512] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.418935] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.420038] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62914) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 527.421238] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager.update_available_resource {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.708680] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352316, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.026145} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.708930] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] File moved {{(pid=62914) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 527.709198] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Cleaning up location [datastore2] vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116 {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 527.709371] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Deleting the datastore file [datastore2] vmware_temp/d49cdfbc-4039-4f6b-8ed0-6356dab8e116 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 527.709613] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c381d958-c546-4c05-a1b4-f1140485b1e4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.719521] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 527.719521] env[62914]: value = "task-1352317" [ 527.719521] env[62914]: _type = "Task" [ 527.719521] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.730878] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352317, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.769094] env[62914]: ERROR nova.scheduler.client.report [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [req-552abc9a-f855-417c-8871-27226b6170ad] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8e1b3185-5ae3-4780-aee8-c507b157ad13. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-552abc9a-f855-417c-8871-27226b6170ad"}]} [ 527.787123] env[62914]: DEBUG nova.scheduler.client.report [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Refreshing inventories for resource provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 527.804025] env[62914]: DEBUG nova.scheduler.client.report [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Updating ProviderTree inventory for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 527.804025] env[62914]: DEBUG nova.compute.provider_tree [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Updating inventory in ProviderTree for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 527.813790] env[62914]: DEBUG nova.scheduler.client.report [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Refreshing aggregate associations for resource provider 8e1b3185-5ae3-4780-aee8-c507b157ad13, aggregates: None {{(pid=62914) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 527.842460] env[62914]: DEBUG nova.scheduler.client.report [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Refreshing trait associations for resource provider 8e1b3185-5ae3-4780-aee8-c507b157ad13, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62914) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 527.877193] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Releasing lock "refresh_cache-33f4159e-4181-4345-a716-0dfc409cee78" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.877583] env[62914]: DEBUG nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 527.877770] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 527.878236] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ccfcd3c-ee07-4c02-836d-3b08af75f2a9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.887269] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533da974-8065-40d6-ad50-3530534d6fff {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.901062] env[62914]: DEBUG nova.network.neutron [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 527.910564] env[62914]: DEBUG nova.compute.manager [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 527.918895] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 33f4159e-4181-4345-a716-0dfc409cee78 could not be found. [ 527.919134] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 527.919314] env[62914]: INFO nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Took 0.04 seconds to destroy the instance on the hypervisor. [ 527.921319] env[62914]: DEBUG oslo.service.loopingcall [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 527.921319] env[62914]: DEBUG nova.compute.manager [-] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 527.921319] env[62914]: DEBUG nova.network.neutron [-] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 527.923849] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.961503] env[62914]: DEBUG nova.network.neutron [-] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 527.967717] env[62914]: DEBUG nova.virt.hardware [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 527.967956] env[62914]: DEBUG nova.virt.hardware [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 527.968118] env[62914]: DEBUG nova.virt.hardware [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 527.969017] env[62914]: DEBUG nova.virt.hardware [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 527.969239] env[62914]: DEBUG nova.virt.hardware [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 527.970237] env[62914]: DEBUG nova.virt.hardware [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 527.970237] env[62914]: DEBUG nova.virt.hardware [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 527.970237] env[62914]: DEBUG nova.virt.hardware [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 527.970237] env[62914]: DEBUG nova.virt.hardware [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 527.970237] env[62914]: DEBUG nova.virt.hardware [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 527.970451] env[62914]: DEBUG nova.virt.hardware [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 527.974073] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f6dd761-ecd4-4194-8ec6-709036d8110d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.981781] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f1185f3-bf40-42d0-88b2-4f4b0ef271b2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.997685] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Instance VIF info [] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 528.003227] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Creating folder: Project (9e9958977815476da50cbe18e0a67495). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 528.006033] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-642f96fb-41bc-442a-bb8b-005da9143e0e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.019056] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Created folder: Project (9e9958977815476da50cbe18e0a67495) in parent group-v288131. [ 528.019056] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Creating folder: Instances. Parent ref: group-v288139. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 528.020458] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1b025e33-fa72-4533-820c-7a28270e269b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.037430] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Created folder: Instances in parent group-v288139. [ 528.037700] env[62914]: DEBUG oslo.service.loopingcall [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 528.037886] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 528.038115] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34df35b1-2d94-432d-911c-b5543a4ecd6b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.063477] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 528.063477] env[62914]: value = "task-1352320" [ 528.063477] env[62914]: _type = "Task" [ 528.063477] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.071802] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352320, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.102473] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4615f592-623f-45b2-969a-096fcacaa1fa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.115156] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-164acdd8-6177-4ed6-b045-fe6030c2d936 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.159700] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2586f3-7062-4203-9859-f4acac189a66 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.166387] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Acquiring lock "a8761350-3044-4822-8221-ed108e21cd62" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.167217] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Lock "a8761350-3044-4822-8221-ed108e21cd62" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.181060] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b891d5df-f6cb-46dc-8114-6f00735f849b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.202764] env[62914]: DEBUG nova.compute.provider_tree [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Updating inventory in ProviderTree for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 528.231603] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352317, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.045762} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.232558] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 528.233376] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99735370-ac5c-4bd2-bd58-b4ba0ce74448 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.245022] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 528.245022] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5248ee71-91b2-04fd-628e-984823dce25e" [ 528.245022] env[62914]: _type = "Task" [ 528.245022] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.247867] env[62914]: DEBUG nova.network.neutron [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.256946] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5248ee71-91b2-04fd-628e-984823dce25e, 'name': SearchDatastore_Task, 'duration_secs': 0.010593} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.257218] env[62914]: DEBUG oslo_concurrency.lockutils [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.257471] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417/3786cd93-1a50-4997-8894-840afa1c8417.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 528.258069] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ebb4be7e-6522-46d4-b227-933adc3d6097 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.268319] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 528.268319] env[62914]: value = "task-1352322" [ 528.268319] env[62914]: _type = "Task" [ 528.268319] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.284181] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352322, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.461148] env[62914]: DEBUG nova.network.neutron [-] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.582682] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352320, 'name': CreateVM_Task, 'duration_secs': 0.318313} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.582903] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 528.584072] env[62914]: DEBUG oslo_vmware.service [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e820913f-e2bd-4e82-a02d-a7deecdda8af {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.594909] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.595252] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.595692] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 528.595990] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45e93e32-e4ad-41c0-9b0d-a9986ddafd39 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.606889] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for the task: (returnval){ [ 528.606889] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ee85cb-4260-9b15-c1ef-8d994de2a62c" [ 528.606889] env[62914]: _type = "Task" [ 528.606889] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.615974] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ee85cb-4260-9b15-c1ef-8d994de2a62c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.679296] env[62914]: DEBUG nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 528.753387] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] Releasing lock "refresh_cache-3db8507b-52b8-4eeb-b801-56e817c3b728" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.753387] env[62914]: DEBUG nova.compute.manager [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Received event network-vif-deleted-ade1c8df-fe65-4c76-9615-8490472a1ed7 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 528.753387] env[62914]: DEBUG nova.compute.manager [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Received event network-changed-44ec2ba3-6f10-4fff-8e78-9872e8b8c0a0 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 528.753387] env[62914]: DEBUG nova.compute.manager [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Refreshing instance network info cache due to event network-changed-44ec2ba3-6f10-4fff-8e78-9872e8b8c0a0. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 528.753387] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] Acquiring lock "refresh_cache-33f4159e-4181-4345-a716-0dfc409cee78" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.753576] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] Acquired lock "refresh_cache-33f4159e-4181-4345-a716-0dfc409cee78" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.753576] env[62914]: DEBUG nova.network.neutron [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Refreshing network info cache for port 44ec2ba3-6f10-4fff-8e78-9872e8b8c0a0 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 528.790575] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352322, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.792100] env[62914]: DEBUG nova.scheduler.client.report [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Updated inventory for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 with generation 14 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 528.792360] env[62914]: DEBUG nova.compute.provider_tree [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Updating resource provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 generation from 14 to 15 during operation: update_inventory {{(pid=62914) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 528.792710] env[62914]: DEBUG nova.compute.provider_tree [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Updating inventory in ProviderTree for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 528.967062] env[62914]: INFO nova.compute.manager [-] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Took 1.05 seconds to deallocate network for instance. [ 528.972722] env[62914]: DEBUG nova.compute.claims [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 528.972834] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.120716] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.121013] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 529.121267] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.121413] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.122021] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 529.122278] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d719d49c-d2b5-47dd-9a04-c20012ca6019 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.132579] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 529.133061] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 529.133831] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705332a8-cbf7-46e1-bed8-7665df889e9e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.141063] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c5f192a-5974-4a6f-bb00-918065a985c4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.147761] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for the task: (returnval){ [ 529.147761] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f46817-596b-1a8d-12e6-0db74f98e2ec" [ 529.147761] env[62914]: _type = "Task" [ 529.147761] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.163271] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Preparing fetch location {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 529.163413] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Creating directory with path [datastore1] vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a/6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 529.163963] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4aefda6a-19fa-4078-9f96-b6c57c577e28 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.178506] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Created directory with path [datastore1] vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a/6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 529.178722] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Fetch image to [datastore1] vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 529.178900] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Downloading image file data 6c21e11e-da68-48a7-9929-737212ce98cf to [datastore1] vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk on the data store datastore1 {{(pid=62914) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 529.179901] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a7b119-d805-4722-bf14-9db7518a4c21 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.191534] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a619cf51-f0f5-425b-9ccd-0e2310019c4f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.202945] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dda5b29-3673-4c8e-bc26-6be997179317 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.245721] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.245995] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ce4f08-f0a9-4e3e-aafd-db7d8d57be40 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.254120] env[62914]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-36e75e55-5c0d-4884-9040-cc823bf36247 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.295725] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352322, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521354} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.296971] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417/3786cd93-1a50-4997-8894-840afa1c8417.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 529.297229] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 529.299727] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Downloading image file data 6c21e11e-da68-48a7-9929-737212ce98cf to the data store datastore1 {{(pid=62914) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 529.301936] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e955a208-11b0-4321-b331-0a2780c9ac04 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.307603] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.439s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.308117] env[62914]: DEBUG nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 529.312771] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.896s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.315398] env[62914]: INFO nova.compute.claims [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.327715] env[62914]: DEBUG nova.network.neutron [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 529.332392] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 529.332392] env[62914]: value = "task-1352323" [ 529.332392] env[62914]: _type = "Task" [ 529.332392] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.347970] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352323, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.386348] env[62914]: DEBUG oslo_vmware.rw_handles [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62914) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 529.723391] env[62914]: DEBUG nova.network.neutron [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.821329] env[62914]: DEBUG nova.compute.utils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 529.832090] env[62914]: DEBUG nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 529.832339] env[62914]: DEBUG nova.network.neutron [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 529.847727] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352323, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070026} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.848352] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 529.849405] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-297b8d43-8bcb-4e85-b5d3-b51afa35fff3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.887557] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417/3786cd93-1a50-4997-8894-840afa1c8417.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 529.891649] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-346a88c6-4853-431e-9445-069545cce9a0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.919478] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 529.919478] env[62914]: value = "task-1352324" [ 529.919478] env[62914]: _type = "Task" [ 529.919478] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.931871] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352324, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.995819] env[62914]: DEBUG nova.policy [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5aa76264662243279f141cf4074b09ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42f0d7b082f0488bb6cf4e4599e20c92', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 530.059837] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Acquiring lock "856b3816-d4f8-4c9e-b74c-ea15d67976ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.060499] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Lock "856b3816-d4f8-4c9e-b74c-ea15d67976ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.081232] env[62914]: DEBUG oslo_vmware.rw_handles [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Completed reading data from the image iterator. {{(pid=62914) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 530.081572] env[62914]: DEBUG oslo_vmware.rw_handles [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62914) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 530.143515] env[62914]: DEBUG nova.compute.manager [req-0b8b33c3-04eb-4186-9f88-e5ee3bdd495a req-f0d75dda-b116-4c66-9330-c0e4414413f0 service nova] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Received event network-vif-deleted-44ec2ba3-6f10-4fff-8e78-9872e8b8c0a0 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 530.227309] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b483896-58a3-4aaa-b014-5bd966e4efca req-309d146a-e27a-4c37-9e5b-185a30961ca6 service nova] Releasing lock "refresh_cache-33f4159e-4181-4345-a716-0dfc409cee78" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.230163] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Downloaded image file data 6c21e11e-da68-48a7-9929-737212ce98cf to vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk on the data store datastore1 {{(pid=62914) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 530.232471] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Caching image {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 530.232829] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Copying Virtual Disk [datastore1] vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk to [datastore1] vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 530.233171] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3fe4363f-dbc0-4fe8-99ca-ffd3020b3e19 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.242180] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for the task: (returnval){ [ 530.242180] env[62914]: value = "task-1352325" [ 530.242180] env[62914]: _type = "Task" [ 530.242180] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.254173] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352325, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.324304] env[62914]: DEBUG nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 530.434354] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352324, 'name': ReconfigVM_Task, 'duration_secs': 0.311298} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.434644] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417/3786cd93-1a50-4997-8894-840afa1c8417.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 530.435342] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4a08d6b7-e8bf-4a46-a4ea-088199167d1f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.450860] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 530.450860] env[62914]: value = "task-1352327" [ 530.450860] env[62914]: _type = "Task" [ 530.450860] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.459785] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352327, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.604253] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b73dc0a4-d341-49e6-a318-a8253c55406e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.610865] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ec2661-14e7-402f-82c5-82c23735dac0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.661796] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d96029cc-eb23-464e-8d26-0c06669a02ae {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.672766] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-905a3497-6bc1-4c7f-a63d-6589484c9bec {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.689195] env[62914]: DEBUG nova.compute.provider_tree [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 530.757434] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352325, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.964926] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352327, 'name': Rename_Task, 'duration_secs': 0.188399} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.965431] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 530.965647] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-23db74bd-022d-4597-bfc7-4c0cc882ea31 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.977030] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 530.977030] env[62914]: value = "task-1352328" [ 530.977030] env[62914]: _type = "Task" [ 530.977030] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.985912] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352328, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.193141] env[62914]: DEBUG nova.scheduler.client.report [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.261015] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352325, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.679278} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.261248] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Copied Virtual Disk [datastore1] vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk to [datastore1] vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 531.262621] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Deleting the datastore file [datastore1] vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a/6c21e11e-da68-48a7-9929-737212ce98cf/tmp-sparse.vmdk {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 531.262621] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c99bcdba-85b3-48d0-b877-900c8bb0fd22 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.273084] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for the task: (returnval){ [ 531.273084] env[62914]: value = "task-1352329" [ 531.273084] env[62914]: _type = "Task" [ 531.273084] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.285234] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352329, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.337893] env[62914]: DEBUG nova.network.neutron [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Successfully created port: 42db1776-19ce-4c94-b288-581cad28a7f9 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 531.343140] env[62914]: DEBUG nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 531.383860] env[62914]: DEBUG nova.virt.hardware [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 531.384201] env[62914]: DEBUG nova.virt.hardware [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 531.384453] env[62914]: DEBUG nova.virt.hardware [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.384784] env[62914]: DEBUG nova.virt.hardware [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 531.384917] env[62914]: DEBUG nova.virt.hardware [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.384976] env[62914]: DEBUG nova.virt.hardware [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 531.385304] env[62914]: DEBUG nova.virt.hardware [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 531.385531] env[62914]: DEBUG nova.virt.hardware [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 531.385713] env[62914]: DEBUG nova.virt.hardware [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 531.385928] env[62914]: DEBUG nova.virt.hardware [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 531.386124] env[62914]: DEBUG nova.virt.hardware [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 531.386985] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4618b856-55d5-4705-b7e9-3b2e6fed11dc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.399272] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57bdff25-8f8c-4a4e-9443-2bf04b0e00ad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.490858] env[62914]: DEBUG oslo_vmware.api [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352328, 'name': PowerOnVM_Task, 'duration_secs': 0.499625} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.491192] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 531.491412] env[62914]: INFO nova.compute.manager [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Took 8.31 seconds to spawn the instance on the hypervisor. [ 531.492060] env[62914]: DEBUG nova.compute.manager [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 531.493272] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36dfa2a6-998e-4603-84d9-95baba54f5de {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.702253] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.702786] env[62914]: DEBUG nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 531.709878] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.432s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.711169] env[62914]: INFO nova.compute.claims [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 531.787766] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352329, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027202} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.788314] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 531.788314] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Moving file from [datastore1] vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a/6c21e11e-da68-48a7-9929-737212ce98cf to [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf. {{(pid=62914) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 531.788374] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-e5a5e34a-966d-468e-be84-c207c9d0d729 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.797663] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for the task: (returnval){ [ 531.797663] env[62914]: value = "task-1352330" [ 531.797663] env[62914]: _type = "Task" [ 531.797663] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.809387] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352330, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.873563] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Acquiring lock "889af877-29c4-4c06-9995-49fb15b21a55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.873797] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Lock "889af877-29c4-4c06-9995-49fb15b21a55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.016458] env[62914]: INFO nova.compute.manager [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Took 15.20 seconds to build instance. [ 532.210400] env[62914]: DEBUG nova.compute.utils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.212986] env[62914]: DEBUG nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 532.213168] env[62914]: DEBUG nova.network.neutron [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 532.312458] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352330, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.029807} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.312458] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] File moved {{(pid=62914) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 532.312687] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Cleaning up location [datastore1] vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 532.313249] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Deleting the datastore file [datastore1] vmware_temp/93c71fbc-1998-4157-8c10-087de94e9d4a {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 532.313537] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a426802-054e-4cd9-a725-5dba67fc9a39 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.322824] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for the task: (returnval){ [ 532.322824] env[62914]: value = "task-1352331" [ 532.322824] env[62914]: _type = "Task" [ 532.322824] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.336196] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352331, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.416737] env[62914]: DEBUG nova.policy [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7c3e1664a6da4ee6bb53860ad83a7087', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0497bf6570de478f8b92829e031be5a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 532.520020] env[62914]: DEBUG oslo_concurrency.lockutils [None req-54d2edfb-8840-4ef2-a08c-f0d00ecea5cc tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Lock "3786cd93-1a50-4997-8894-840afa1c8417" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.714s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.713012] env[62914]: DEBUG nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 532.834922] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352331, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.028922} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.837105] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 532.840424] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cda8fa4-b1a3-4a28-be29-b71cd7aafa38 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.848796] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for the task: (returnval){ [ 532.848796] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a31dfa-70ca-f5da-97c3-bdf5880d886d" [ 532.848796] env[62914]: _type = "Task" [ 532.848796] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.862145] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a31dfa-70ca-f5da-97c3-bdf5880d886d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.932040] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a7f5885-87a9-4d88-84bf-928d50a2bd93 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.950659] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22b0f1a-6934-4748-ae08-3f11e5f76e86 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.991278] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d95591-118d-4b20-9348-595370ceb397 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.001197] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c49c46c-280a-477c-94a1-4329bfbf86b4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.016809] env[62914]: DEBUG nova.compute.provider_tree [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 533.022112] env[62914]: DEBUG nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 533.230875] env[62914]: DEBUG nova.network.neutron [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Successfully created port: b615410e-0168-495f-b353-68103e3c6134 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 533.361850] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a31dfa-70ca-f5da-97c3-bdf5880d886d, 'name': SearchDatastore_Task, 'duration_secs': 0.011164} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.362204] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.362609] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] d5c90d13-41bc-4cca-abf3-5a5b13171a6f/d5c90d13-41bc-4cca-abf3-5a5b13171a6f.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 533.362772] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-68ae5b02-dd36-4b9d-831c-94a03c6145d4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.375439] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for the task: (returnval){ [ 533.375439] env[62914]: value = "task-1352332" [ 533.375439] env[62914]: _type = "Task" [ 533.375439] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.394058] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352332, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.523728] env[62914]: DEBUG nova.scheduler.client.report [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 533.557565] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.728081] env[62914]: DEBUG nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 533.763896] env[62914]: DEBUG nova.virt.hardware [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 533.764356] env[62914]: DEBUG nova.virt.hardware [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 533.764618] env[62914]: DEBUG nova.virt.hardware [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 533.764997] env[62914]: DEBUG nova.virt.hardware [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 533.765340] env[62914]: DEBUG nova.virt.hardware [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 533.765592] env[62914]: DEBUG nova.virt.hardware [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 533.767027] env[62914]: DEBUG nova.virt.hardware [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 533.767027] env[62914]: DEBUG nova.virt.hardware [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 533.767027] env[62914]: DEBUG nova.virt.hardware [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 533.767302] env[62914]: DEBUG nova.virt.hardware [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 533.768233] env[62914]: DEBUG nova.virt.hardware [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 533.771016] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b48088a-247c-4c25-b771-6228548636bb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.782375] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05347979-9f4f-4526-b5dc-25451a198879 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.891306] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352332, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.029135] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.320s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.031846] env[62914]: DEBUG nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 534.037875] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.584s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.246904] env[62914]: ERROR nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 42db1776-19ce-4c94-b288-581cad28a7f9, please check neutron logs for more information. [ 534.246904] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 534.246904] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.246904] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 534.246904] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.246904] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 534.246904] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.246904] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 534.246904] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.246904] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 534.246904] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.246904] env[62914]: ERROR nova.compute.manager raise self.value [ 534.246904] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.246904] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 534.246904] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.246904] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 534.247453] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.247453] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 534.247453] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 42db1776-19ce-4c94-b288-581cad28a7f9, please check neutron logs for more information. [ 534.247453] env[62914]: ERROR nova.compute.manager [ 534.247453] env[62914]: Traceback (most recent call last): [ 534.247453] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 534.247453] env[62914]: listener.cb(fileno) [ 534.247453] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.247453] env[62914]: result = function(*args, **kwargs) [ 534.247453] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.247453] env[62914]: return func(*args, **kwargs) [ 534.247453] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.247453] env[62914]: raise e [ 534.247453] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.247453] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 534.247453] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.247453] env[62914]: created_port_ids = self._update_ports_for_instance( [ 534.247453] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.247453] env[62914]: with excutils.save_and_reraise_exception(): [ 534.247453] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.247453] env[62914]: self.force_reraise() [ 534.247453] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.247453] env[62914]: raise self.value [ 534.247453] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.247453] env[62914]: updated_port = self._update_port( [ 534.247453] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.247453] env[62914]: _ensure_no_port_binding_failure(port) [ 534.247453] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.247453] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 534.248694] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 42db1776-19ce-4c94-b288-581cad28a7f9, please check neutron logs for more information. [ 534.248694] env[62914]: Removing descriptor: 15 [ 534.248694] env[62914]: ERROR nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 42db1776-19ce-4c94-b288-581cad28a7f9, please check neutron logs for more information. [ 534.248694] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Traceback (most recent call last): [ 534.248694] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 534.248694] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] yield resources [ 534.248694] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 534.248694] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] self.driver.spawn(context, instance, image_meta, [ 534.248694] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 534.248694] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 534.248694] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 534.248694] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] vm_ref = self.build_virtual_machine(instance, [ 534.249079] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 534.249079] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 534.249079] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 534.249079] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] for vif in network_info: [ 534.249079] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 534.249079] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] return self._sync_wrapper(fn, *args, **kwargs) [ 534.249079] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 534.249079] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] self.wait() [ 534.249079] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 534.249079] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] self[:] = self._gt.wait() [ 534.249079] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 534.249079] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] return self._exit_event.wait() [ 534.249079] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 534.249515] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] result = hub.switch() [ 534.249515] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 534.249515] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] return self.greenlet.switch() [ 534.249515] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.249515] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] result = function(*args, **kwargs) [ 534.249515] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.249515] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] return func(*args, **kwargs) [ 534.249515] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.249515] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] raise e [ 534.249515] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.249515] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] nwinfo = self.network_api.allocate_for_instance( [ 534.249515] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.249515] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] created_port_ids = self._update_ports_for_instance( [ 534.249845] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.249845] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] with excutils.save_and_reraise_exception(): [ 534.249845] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.249845] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] self.force_reraise() [ 534.249845] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.249845] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] raise self.value [ 534.249845] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.249845] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] updated_port = self._update_port( [ 534.249845] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.249845] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] _ensure_no_port_binding_failure(port) [ 534.249845] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.249845] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] raise exception.PortBindingFailed(port_id=port['id']) [ 534.250141] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] nova.exception.PortBindingFailed: Binding failed for port 42db1776-19ce-4c94-b288-581cad28a7f9, please check neutron logs for more information. [ 534.250141] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] [ 534.250141] env[62914]: INFO nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Terminating instance [ 534.254851] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Acquiring lock "refresh_cache-a7ee1dbd-73d2-46be-bf43-008eebcd64e9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.254851] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Acquired lock "refresh_cache-a7ee1dbd-73d2-46be-bf43-008eebcd64e9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.254851] env[62914]: DEBUG nova.network.neutron [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 534.387874] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352332, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.62036} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.389202] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] d5c90d13-41bc-4cca-abf3-5a5b13171a6f/d5c90d13-41bc-4cca-abf3-5a5b13171a6f.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 534.393176] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 534.393176] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-37a432ea-c114-4fbc-8478-d0f52c036a4e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.405585] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for the task: (returnval){ [ 534.405585] env[62914]: value = "task-1352333" [ 534.405585] env[62914]: _type = "Task" [ 534.405585] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.418255] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352333, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.548687] env[62914]: DEBUG nova.compute.utils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 534.551848] env[62914]: DEBUG nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 534.551848] env[62914]: DEBUG nova.network.neutron [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 534.729467] env[62914]: DEBUG nova.policy [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce06463e715b4975bc43b14be64ef9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '619be7b3bf1445b68ba7adefc98f8782', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 534.763710] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f217eac7-8389-4323-ba4c-9ab376ee0bcc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.775240] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55473bac-810a-46af-83fe-5e478964fdff {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.808144] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9756995f-45ce-4563-8ce8-3ec4d0b77085 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.812678] env[62914]: DEBUG nova.network.neutron [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.818291] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f068a2f-2d94-4e9d-b023-d12009699f4f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.833579] env[62914]: DEBUG nova.compute.provider_tree [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 534.918718] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352333, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.229868} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.918957] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 534.920025] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e001b945-b0ca-4d03-8ade-2ca4ba1b3d60 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.947713] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Reconfiguring VM instance instance-00000006 to attach disk [datastore1] d5c90d13-41bc-4cca-abf3-5a5b13171a6f/d5c90d13-41bc-4cca-abf3-5a5b13171a6f.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 534.948238] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e382eb9-6163-4296-8b8c-dc667b1cbaf9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.973425] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for the task: (returnval){ [ 534.973425] env[62914]: value = "task-1352334" [ 534.973425] env[62914]: _type = "Task" [ 534.973425] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.984075] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352334, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.055643] env[62914]: DEBUG nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 535.313978] env[62914]: DEBUG nova.network.neutron [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.340904] env[62914]: DEBUG nova.scheduler.client.report [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 535.489057] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352334, 'name': ReconfigVM_Task, 'duration_secs': 0.30647} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.489057] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Reconfigured VM instance instance-00000006 to attach disk [datastore1] d5c90d13-41bc-4cca-abf3-5a5b13171a6f/d5c90d13-41bc-4cca-abf3-5a5b13171a6f.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 535.489057] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-32778a48-1468-4136-88ea-3740d6780227 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.499848] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for the task: (returnval){ [ 535.499848] env[62914]: value = "task-1352335" [ 535.499848] env[62914]: _type = "Task" [ 535.499848] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.516749] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352335, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.769368] env[62914]: DEBUG nova.compute.manager [req-ad138fdd-f29a-4ca4-9b69-b280d6d47d81 req-5999af3f-74f6-4dc8-af31-d632cf002754 service nova] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Received event network-changed-42db1776-19ce-4c94-b288-581cad28a7f9 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 535.769549] env[62914]: DEBUG nova.compute.manager [req-ad138fdd-f29a-4ca4-9b69-b280d6d47d81 req-5999af3f-74f6-4dc8-af31-d632cf002754 service nova] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Refreshing instance network info cache due to event network-changed-42db1776-19ce-4c94-b288-581cad28a7f9. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 535.769753] env[62914]: DEBUG oslo_concurrency.lockutils [req-ad138fdd-f29a-4ca4-9b69-b280d6d47d81 req-5999af3f-74f6-4dc8-af31-d632cf002754 service nova] Acquiring lock "refresh_cache-a7ee1dbd-73d2-46be-bf43-008eebcd64e9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.816540] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Releasing lock "refresh_cache-a7ee1dbd-73d2-46be-bf43-008eebcd64e9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.817560] env[62914]: DEBUG nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 535.817849] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 535.818598] env[62914]: DEBUG oslo_concurrency.lockutils [req-ad138fdd-f29a-4ca4-9b69-b280d6d47d81 req-5999af3f-74f6-4dc8-af31-d632cf002754 service nova] Acquired lock "refresh_cache-a7ee1dbd-73d2-46be-bf43-008eebcd64e9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.818598] env[62914]: DEBUG nova.network.neutron [req-ad138fdd-f29a-4ca4-9b69-b280d6d47d81 req-5999af3f-74f6-4dc8-af31-d632cf002754 service nova] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Refreshing network info cache for port 42db1776-19ce-4c94-b288-581cad28a7f9 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 535.819658] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-749a2386-3c97-49bc-a1a9-16a7d0eec814 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.830788] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0773aa46-ede3-40ac-b71a-812def3e2e5b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.847212] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.810s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.847918] env[62914]: ERROR nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6fde6d50-bd0f-4768-911b-5b20080bdfa3, please check neutron logs for more information. [ 535.847918] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Traceback (most recent call last): [ 535.847918] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 535.847918] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] self.driver.spawn(context, instance, image_meta, [ 535.847918] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 535.847918] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 535.847918] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 535.847918] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] vm_ref = self.build_virtual_machine(instance, [ 535.847918] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 535.847918] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] vif_infos = vmwarevif.get_vif_info(self._session, [ 535.847918] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 535.848288] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] for vif in network_info: [ 535.848288] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 535.848288] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] return self._sync_wrapper(fn, *args, **kwargs) [ 535.848288] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 535.848288] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] self.wait() [ 535.848288] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 535.848288] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] self[:] = self._gt.wait() [ 535.848288] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 535.848288] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] return self._exit_event.wait() [ 535.848288] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 535.848288] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] result = hub.switch() [ 535.848288] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 535.848288] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] return self.greenlet.switch() [ 535.848685] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.848685] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] result = function(*args, **kwargs) [ 535.848685] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 535.848685] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] return func(*args, **kwargs) [ 535.848685] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.848685] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] raise e [ 535.848685] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.848685] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] nwinfo = self.network_api.allocate_for_instance( [ 535.848685] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 535.848685] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] created_port_ids = self._update_ports_for_instance( [ 535.848685] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 535.848685] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] with excutils.save_and_reraise_exception(): [ 535.848685] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.849023] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] self.force_reraise() [ 535.849023] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.849023] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] raise self.value [ 535.849023] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 535.849023] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] updated_port = self._update_port( [ 535.849023] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.849023] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] _ensure_no_port_binding_failure(port) [ 535.849023] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.849023] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] raise exception.PortBindingFailed(port_id=port['id']) [ 535.849023] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] nova.exception.PortBindingFailed: Binding failed for port 6fde6d50-bd0f-4768-911b-5b20080bdfa3, please check neutron logs for more information. [ 535.849023] env[62914]: ERROR nova.compute.manager [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] [ 535.849407] env[62914]: DEBUG nova.compute.utils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Binding failed for port 6fde6d50-bd0f-4768-911b-5b20080bdfa3, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 535.852291] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.167s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.852291] env[62914]: INFO nova.compute.claims [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 535.870169] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a7ee1dbd-73d2-46be-bf43-008eebcd64e9 could not be found. [ 535.870169] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 535.870169] env[62914]: INFO nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 535.870639] env[62914]: DEBUG oslo.service.loopingcall [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.871573] env[62914]: DEBUG nova.compute.manager [-] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.871656] env[62914]: DEBUG nova.network.neutron [-] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 535.879989] env[62914]: DEBUG nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Build of instance b100dbb2-7e2e-43da-bb5f-e31c6676d1bd was re-scheduled: Binding failed for port 6fde6d50-bd0f-4768-911b-5b20080bdfa3, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 535.880496] env[62914]: DEBUG nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 535.880739] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Acquiring lock "refresh_cache-b100dbb2-7e2e-43da-bb5f-e31c6676d1bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.880891] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Acquired lock "refresh_cache-b100dbb2-7e2e-43da-bb5f-e31c6676d1bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.881063] env[62914]: DEBUG nova.network.neutron [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 535.947379] env[62914]: DEBUG nova.network.neutron [-] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.954131] env[62914]: DEBUG nova.network.neutron [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Successfully created port: 0451d3e7-8933-4ae7-ad82-9ffde652798f {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 536.015774] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352335, 'name': Rename_Task, 'duration_secs': 0.147434} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.016086] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 536.017579] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-91e11e5e-a502-4749-afa5-e3f73629d06b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.023793] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for the task: (returnval){ [ 536.023793] env[62914]: value = "task-1352336" [ 536.023793] env[62914]: _type = "Task" [ 536.023793] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.034469] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352336, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.055324] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Acquiring lock "4981a04f-98e6-474b-850a-634d977d9e42" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.055787] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Lock "4981a04f-98e6-474b-850a-634d977d9e42" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.072513] env[62914]: DEBUG nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 536.091350] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Acquiring lock "f1cee1cb-aceb-47e8-b2da-b94563f922e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.091854] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Lock "f1cee1cb-aceb-47e8-b2da-b94563f922e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.111189] env[62914]: DEBUG nova.virt.hardware [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 536.112463] env[62914]: DEBUG nova.virt.hardware [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 536.112463] env[62914]: DEBUG nova.virt.hardware [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.112463] env[62914]: DEBUG nova.virt.hardware [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 536.112463] env[62914]: DEBUG nova.virt.hardware [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.112463] env[62914]: DEBUG nova.virt.hardware [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 536.113726] env[62914]: DEBUG nova.virt.hardware [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 536.113726] env[62914]: DEBUG nova.virt.hardware [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 536.113726] env[62914]: DEBUG nova.virt.hardware [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 536.113726] env[62914]: DEBUG nova.virt.hardware [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 536.113726] env[62914]: DEBUG nova.virt.hardware [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 536.117128] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c402a9d-0d8a-4d38-8f92-d76220c93e1d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.129327] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b7b2f8e-9771-40a1-84a9-d288ddfe7b4c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.204306] env[62914]: INFO nova.compute.manager [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Rebuilding instance [ 536.267420] env[62914]: DEBUG nova.compute.manager [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 536.268364] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-508f40c6-f912-4934-a4aa-2e95c803c362 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.433513] env[62914]: DEBUG nova.network.neutron [req-ad138fdd-f29a-4ca4-9b69-b280d6d47d81 req-5999af3f-74f6-4dc8-af31-d632cf002754 service nova] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.438970] env[62914]: DEBUG nova.network.neutron [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.454565] env[62914]: DEBUG nova.network.neutron [-] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.538812] env[62914]: DEBUG oslo_vmware.api [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352336, 'name': PowerOnVM_Task, 'duration_secs': 0.465013} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.538812] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 536.538812] env[62914]: INFO nova.compute.manager [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Took 8.63 seconds to spawn the instance on the hypervisor. [ 536.538812] env[62914]: DEBUG nova.compute.manager [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 536.538946] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ffabc6-2897-4e75-8d62-bbbecd7a0209 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.739507] env[62914]: DEBUG nova.network.neutron [req-ad138fdd-f29a-4ca4-9b69-b280d6d47d81 req-5999af3f-74f6-4dc8-af31-d632cf002754 service nova] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.781920] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 536.782224] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87375e70-e369-4d6a-ad65-b1036479d71e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.790728] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 536.790728] env[62914]: value = "task-1352337" [ 536.790728] env[62914]: _type = "Task" [ 536.790728] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.800580] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352337, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.855072] env[62914]: DEBUG nova.network.neutron [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.959644] env[62914]: INFO nova.compute.manager [-] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Took 1.09 seconds to deallocate network for instance. [ 536.962131] env[62914]: DEBUG nova.compute.claims [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 536.962400] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.057802] env[62914]: INFO nova.compute.manager [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Took 18.21 seconds to build instance. [ 537.075149] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bf51bb-b6d3-499c-a42c-bc8ea3525559 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.089021] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bef7b99-d3ea-4797-9df4-7b36628402e5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.122799] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83fe7603-ceac-464d-81bc-e6b2d7018640 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.131956] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccfbd47f-aaa4-4599-b6c7-e0aa5bcb3dd2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.146023] env[62914]: DEBUG nova.compute.provider_tree [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 537.243304] env[62914]: ERROR nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b615410e-0168-495f-b353-68103e3c6134, please check neutron logs for more information. [ 537.243304] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 537.243304] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.243304] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 537.243304] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.243304] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 537.243304] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.243304] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 537.243304] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.243304] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 537.243304] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.243304] env[62914]: ERROR nova.compute.manager raise self.value [ 537.243304] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.243304] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 537.243304] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.243304] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 537.243906] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.243906] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 537.243906] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b615410e-0168-495f-b353-68103e3c6134, please check neutron logs for more information. [ 537.243906] env[62914]: ERROR nova.compute.manager [ 537.243906] env[62914]: Traceback (most recent call last): [ 537.243906] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 537.243906] env[62914]: listener.cb(fileno) [ 537.243906] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.243906] env[62914]: result = function(*args, **kwargs) [ 537.243906] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 537.243906] env[62914]: return func(*args, **kwargs) [ 537.243906] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.243906] env[62914]: raise e [ 537.243906] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.243906] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 537.243906] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.243906] env[62914]: created_port_ids = self._update_ports_for_instance( [ 537.243906] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.243906] env[62914]: with excutils.save_and_reraise_exception(): [ 537.243906] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.243906] env[62914]: self.force_reraise() [ 537.243906] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.243906] env[62914]: raise self.value [ 537.243906] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.243906] env[62914]: updated_port = self._update_port( [ 537.243906] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.243906] env[62914]: _ensure_no_port_binding_failure(port) [ 537.243906] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.243906] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 537.244739] env[62914]: nova.exception.PortBindingFailed: Binding failed for port b615410e-0168-495f-b353-68103e3c6134, please check neutron logs for more information. [ 537.244739] env[62914]: Removing descriptor: 19 [ 537.244739] env[62914]: DEBUG oslo_concurrency.lockutils [req-ad138fdd-f29a-4ca4-9b69-b280d6d47d81 req-5999af3f-74f6-4dc8-af31-d632cf002754 service nova] Releasing lock "refresh_cache-a7ee1dbd-73d2-46be-bf43-008eebcd64e9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.244739] env[62914]: ERROR nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b615410e-0168-495f-b353-68103e3c6134, please check neutron logs for more information. [ 537.244739] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Traceback (most recent call last): [ 537.244739] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 537.244739] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] yield resources [ 537.244739] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 537.244739] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] self.driver.spawn(context, instance, image_meta, [ 537.244739] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 537.244739] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.245110] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.245110] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] vm_ref = self.build_virtual_machine(instance, [ 537.245110] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.245110] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.245110] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.245110] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] for vif in network_info: [ 537.245110] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 537.245110] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] return self._sync_wrapper(fn, *args, **kwargs) [ 537.245110] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 537.245110] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] self.wait() [ 537.245110] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 537.245110] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] self[:] = self._gt.wait() [ 537.245110] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.245521] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] return self._exit_event.wait() [ 537.245521] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.245521] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] result = hub.switch() [ 537.245521] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.245521] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] return self.greenlet.switch() [ 537.245521] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.245521] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] result = function(*args, **kwargs) [ 537.245521] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 537.245521] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] return func(*args, **kwargs) [ 537.245521] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.245521] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] raise e [ 537.245521] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.245521] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] nwinfo = self.network_api.allocate_for_instance( [ 537.246148] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.246148] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] created_port_ids = self._update_ports_for_instance( [ 537.246148] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.246148] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] with excutils.save_and_reraise_exception(): [ 537.246148] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.246148] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] self.force_reraise() [ 537.246148] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.246148] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] raise self.value [ 537.246148] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.246148] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] updated_port = self._update_port( [ 537.246148] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.246148] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] _ensure_no_port_binding_failure(port) [ 537.246148] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.246547] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] raise exception.PortBindingFailed(port_id=port['id']) [ 537.246547] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] nova.exception.PortBindingFailed: Binding failed for port b615410e-0168-495f-b353-68103e3c6134, please check neutron logs for more information. [ 537.246547] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] [ 537.246547] env[62914]: INFO nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Terminating instance [ 537.246823] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Acquiring lock "refresh_cache-c78502a7-266d-49cf-9d50-cc89ed266647" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.247197] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Acquired lock "refresh_cache-c78502a7-266d-49cf-9d50-cc89ed266647" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.247437] env[62914]: DEBUG nova.network.neutron [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 537.301539] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352337, 'name': PowerOffVM_Task, 'duration_secs': 0.120687} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.301776] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 537.301994] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 537.302767] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49343aa2-527d-47e7-a2f7-00cd349d3062 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.311536] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 537.311752] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8d5ac681-13ca-4953-8c7b-23b08c9f0a37 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.336914] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 537.337171] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 537.338030] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Deleting the datastore file [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 537.338030] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ebbd60d-fefc-4fd5-b226-39136b439954 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.346025] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 537.346025] env[62914]: value = "task-1352339" [ 537.346025] env[62914]: _type = "Task" [ 537.346025] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.355194] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352339, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.357784] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Releasing lock "refresh_cache-b100dbb2-7e2e-43da-bb5f-e31c6676d1bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.357959] env[62914]: DEBUG nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 537.358128] env[62914]: DEBUG nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 537.358290] env[62914]: DEBUG nova.network.neutron [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 537.395455] env[62914]: DEBUG nova.network.neutron [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.560282] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf6073c3-eb15-4a69-9d0a-53a6444453ae tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Lock "d5c90d13-41bc-4cca-abf3-5a5b13171a6f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.727s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.650119] env[62914]: DEBUG nova.scheduler.client.report [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 537.784597] env[62914]: DEBUG nova.network.neutron [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.858947] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352339, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158322} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.859834] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 537.859834] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 537.859834] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 537.898915] env[62914]: DEBUG nova.network.neutron [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.000748] env[62914]: DEBUG nova.network.neutron [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.065200] env[62914]: DEBUG nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 538.151113] env[62914]: DEBUG nova.compute.manager [req-fcc73225-2d18-4009-9c06-84716d11947c req-18868200-fa16-46f0-9c5d-d0bf0e031bb2 service nova] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Received event network-changed-b615410e-0168-495f-b353-68103e3c6134 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.151363] env[62914]: DEBUG nova.compute.manager [req-fcc73225-2d18-4009-9c06-84716d11947c req-18868200-fa16-46f0-9c5d-d0bf0e031bb2 service nova] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Refreshing instance network info cache due to event network-changed-b615410e-0168-495f-b353-68103e3c6134. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 538.151454] env[62914]: DEBUG oslo_concurrency.lockutils [req-fcc73225-2d18-4009-9c06-84716d11947c req-18868200-fa16-46f0-9c5d-d0bf0e031bb2 service nova] Acquiring lock "refresh_cache-c78502a7-266d-49cf-9d50-cc89ed266647" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.156409] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.306s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.156409] env[62914]: DEBUG nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 538.159010] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.542s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.219280] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Acquiring lock "6ec0c746-2bb8-4387-8b31-b49b0d708614" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.219280] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Lock "6ec0c746-2bb8-4387-8b31-b49b0d708614" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.407840] env[62914]: INFO nova.compute.manager [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] [instance: b100dbb2-7e2e-43da-bb5f-e31c6676d1bd] Took 1.05 seconds to deallocate network for instance. [ 538.504203] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Releasing lock "refresh_cache-c78502a7-266d-49cf-9d50-cc89ed266647" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.505049] env[62914]: DEBUG nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 538.505348] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 538.505678] env[62914]: DEBUG oslo_concurrency.lockutils [req-fcc73225-2d18-4009-9c06-84716d11947c req-18868200-fa16-46f0-9c5d-d0bf0e031bb2 service nova] Acquired lock "refresh_cache-c78502a7-266d-49cf-9d50-cc89ed266647" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.505863] env[62914]: DEBUG nova.network.neutron [req-fcc73225-2d18-4009-9c06-84716d11947c req-18868200-fa16-46f0-9c5d-d0bf0e031bb2 service nova] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Refreshing network info cache for port b615410e-0168-495f-b353-68103e3c6134 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 538.509046] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e64fd261-3c50-4e3f-bcf8-18cf14d88db4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.519831] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a80a42-0988-4a81-a7b2-514287e05e9e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.559758] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c78502a7-266d-49cf-9d50-cc89ed266647 could not be found. [ 538.560524] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 538.560524] env[62914]: INFO nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Took 0.06 seconds to destroy the instance on the hypervisor. [ 538.561115] env[62914]: DEBUG oslo.service.loopingcall [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 538.562678] env[62914]: DEBUG nova.compute.manager [-] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.563072] env[62914]: DEBUG nova.network.neutron [-] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 538.579530] env[62914]: DEBUG nova.compute.manager [req-f24b1084-35a9-400d-bf48-8b5f41525fc7 req-975338d7-0272-4e45-854e-c9233553c14d service nova] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Received event network-vif-deleted-42db1776-19ce-4c94-b288-581cad28a7f9 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.595138] env[62914]: DEBUG nova.network.neutron [-] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.606624] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.666146] env[62914]: DEBUG nova.compute.utils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 538.666146] env[62914]: DEBUG nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 538.666903] env[62914]: DEBUG nova.network.neutron [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 538.810364] env[62914]: DEBUG nova.policy [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7cbf9a3bb65a4b83a484776607fea1fb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd3f0f9be404d4cb898f76f404a0a16a7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 538.912475] env[62914]: DEBUG nova.virt.hardware [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 538.912475] env[62914]: DEBUG nova.virt.hardware [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 538.912475] env[62914]: DEBUG nova.virt.hardware [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 538.912475] env[62914]: DEBUG nova.virt.hardware [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 538.912685] env[62914]: DEBUG nova.virt.hardware [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 538.912685] env[62914]: DEBUG nova.virt.hardware [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 538.912685] env[62914]: DEBUG nova.virt.hardware [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 538.914668] env[62914]: DEBUG nova.virt.hardware [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 538.914668] env[62914]: DEBUG nova.virt.hardware [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 538.914668] env[62914]: DEBUG nova.virt.hardware [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 538.914668] env[62914]: DEBUG nova.virt.hardware [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 538.919479] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee170c66-e2ee-4224-afd2-004a537c226b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.938013] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106c4940-a3b4-4b0a-b716-15e2cccce043 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.960335] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Instance VIF info [] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 538.965980] env[62914]: DEBUG oslo.service.loopingcall [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 538.969026] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 538.969897] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32925b8e-927d-403a-9c72-2e38987bcb1a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.991642] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 538.991642] env[62914]: value = "task-1352340" [ 538.991642] env[62914]: _type = "Task" [ 538.991642] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.001911] env[62914]: DEBUG nova.compute.manager [None req-ea2767b5-e71d-4c36-9e0b-3a2e99ca365c tempest-ServerDiagnosticsV248Test-2004122332 tempest-ServerDiagnosticsV248Test-2004122332-project-admin] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 539.003263] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642b9a80-4904-4055-b747-22020dccd60d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.012030] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352340, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.020244] env[62914]: INFO nova.compute.manager [None req-ea2767b5-e71d-4c36-9e0b-3a2e99ca365c tempest-ServerDiagnosticsV248Test-2004122332 tempest-ServerDiagnosticsV248Test-2004122332-project-admin] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Retrieving diagnostics [ 539.021081] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e935e06-e7fd-40b3-a223-7b0a4b24beb6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.024930] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62186b31-3236-4042-a69c-309fb5cf556a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.064655] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9809f7-7dca-4774-9f81-76a65eb2138f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.072543] env[62914]: DEBUG nova.network.neutron [req-fcc73225-2d18-4009-9c06-84716d11947c req-18868200-fa16-46f0-9c5d-d0bf0e031bb2 service nova] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.100040] env[62914]: DEBUG nova.network.neutron [-] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.102687] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe6a81a-6bba-4087-9175-f33d95b78d9b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.112096] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-518fc21c-e888-4c16-af2b-4ccc039c6cc5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.130654] env[62914]: DEBUG nova.compute.provider_tree [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 539.176149] env[62914]: DEBUG nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 539.455531] env[62914]: DEBUG nova.network.neutron [req-fcc73225-2d18-4009-9c06-84716d11947c req-18868200-fa16-46f0-9c5d-d0bf0e031bb2 service nova] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.459942] env[62914]: INFO nova.scheduler.client.report [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Deleted allocations for instance b100dbb2-7e2e-43da-bb5f-e31c6676d1bd [ 539.511819] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352340, 'name': CreateVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.607615] env[62914]: INFO nova.compute.manager [-] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Took 1.04 seconds to deallocate network for instance. [ 539.610520] env[62914]: DEBUG nova.compute.claims [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 539.611500] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.633880] env[62914]: DEBUG nova.scheduler.client.report [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 539.970358] env[62914]: DEBUG oslo_concurrency.lockutils [req-fcc73225-2d18-4009-9c06-84716d11947c req-18868200-fa16-46f0-9c5d-d0bf0e031bb2 service nova] Releasing lock "refresh_cache-c78502a7-266d-49cf-9d50-cc89ed266647" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.970358] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13edb669-a29b-47cc-84c4-dbb4c288cdf3 tempest-ServerDiagnosticsNegativeTest-582647436 tempest-ServerDiagnosticsNegativeTest-582647436-project-member] Lock "b100dbb2-7e2e-43da-bb5f-e31c6676d1bd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.609s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.010123] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352340, 'name': CreateVM_Task, 'duration_secs': 0.770468} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.011446] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 540.012103] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.012362] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.012772] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 540.013073] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42a1329f-8860-4756-9e29-416e25df844c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.020892] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 540.020892] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c758e8-e45e-06a7-91fa-6942c15db9b5" [ 540.020892] env[62914]: _type = "Task" [ 540.020892] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.032670] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c758e8-e45e-06a7-91fa-6942c15db9b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.138563] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.979s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.138769] env[62914]: ERROR nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ade1c8df-fe65-4c76-9615-8490472a1ed7, please check neutron logs for more information. [ 540.138769] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Traceback (most recent call last): [ 540.138769] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 540.138769] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] self.driver.spawn(context, instance, image_meta, [ 540.138769] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 540.138769] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.138769] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.138769] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] vm_ref = self.build_virtual_machine(instance, [ 540.138769] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.138769] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.138769] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.139117] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] for vif in network_info: [ 540.139117] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.139117] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] return self._sync_wrapper(fn, *args, **kwargs) [ 540.139117] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.139117] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] self.wait() [ 540.139117] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.139117] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] self[:] = self._gt.wait() [ 540.139117] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.139117] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] return self._exit_event.wait() [ 540.139117] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.139117] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] result = hub.switch() [ 540.139117] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.139117] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] return self.greenlet.switch() [ 540.139528] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.139528] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] result = function(*args, **kwargs) [ 540.139528] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 540.139528] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] return func(*args, **kwargs) [ 540.139528] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.139528] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] raise e [ 540.139528] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.139528] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] nwinfo = self.network_api.allocate_for_instance( [ 540.139528] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.139528] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] created_port_ids = self._update_ports_for_instance( [ 540.139528] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.139528] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] with excutils.save_and_reraise_exception(): [ 540.139528] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.139946] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] self.force_reraise() [ 540.139946] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.139946] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] raise self.value [ 540.139946] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.139946] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] updated_port = self._update_port( [ 540.139946] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.139946] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] _ensure_no_port_binding_failure(port) [ 540.139946] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.139946] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] raise exception.PortBindingFailed(port_id=port['id']) [ 540.139946] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] nova.exception.PortBindingFailed: Binding failed for port ade1c8df-fe65-4c76-9615-8490472a1ed7, please check neutron logs for more information. [ 540.139946] env[62914]: ERROR nova.compute.manager [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] [ 540.140559] env[62914]: DEBUG nova.compute.utils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Binding failed for port ade1c8df-fe65-4c76-9615-8490472a1ed7, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 540.140559] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 12.217s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.140656] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.140841] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62914) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 540.141572] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.168s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.144106] env[62914]: DEBUG nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Build of instance 3db8507b-52b8-4eeb-b801-56e817c3b728 was re-scheduled: Binding failed for port ade1c8df-fe65-4c76-9615-8490472a1ed7, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 540.144553] env[62914]: DEBUG nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 540.144764] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Acquiring lock "refresh_cache-3db8507b-52b8-4eeb-b801-56e817c3b728" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.145214] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Acquired lock "refresh_cache-3db8507b-52b8-4eeb-b801-56e817c3b728" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.146987] env[62914]: DEBUG nova.network.neutron [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 540.151381] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4dde8fc-4ed7-4883-bff4-a2ef49a3f627 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.172319] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d54a4b-8805-41ad-b99d-fa171556209d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.190610] env[62914]: DEBUG nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 540.194141] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae01151-7ece-4c49-a3d8-be3b55128583 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.204284] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae5f2b4-600c-4793-8851-aa441972e447 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.247481] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181276MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62914) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 540.247781] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.256358] env[62914]: DEBUG nova.virt.hardware [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 540.256508] env[62914]: DEBUG nova.virt.hardware [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 540.256717] env[62914]: DEBUG nova.virt.hardware [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 540.256803] env[62914]: DEBUG nova.virt.hardware [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 540.257205] env[62914]: DEBUG nova.virt.hardware [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 540.257205] env[62914]: DEBUG nova.virt.hardware [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 540.257314] env[62914]: DEBUG nova.virt.hardware [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 540.257485] env[62914]: DEBUG nova.virt.hardware [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 540.257647] env[62914]: DEBUG nova.virt.hardware [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 540.257800] env[62914]: DEBUG nova.virt.hardware [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 540.257962] env[62914]: DEBUG nova.virt.hardware [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 540.259250] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd980aa-997c-4538-a5c0-4e9e2f630d0c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.268630] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df99dbc-8423-47d1-8687-8f3c7e73bc00 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.472434] env[62914]: DEBUG nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 540.491818] env[62914]: ERROR nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0451d3e7-8933-4ae7-ad82-9ffde652798f, please check neutron logs for more information. [ 540.491818] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 540.491818] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.491818] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 540.491818] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.491818] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 540.491818] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.491818] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 540.491818] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.491818] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 540.491818] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.491818] env[62914]: ERROR nova.compute.manager raise self.value [ 540.491818] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.491818] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 540.491818] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.491818] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 540.492398] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.492398] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 540.492398] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0451d3e7-8933-4ae7-ad82-9ffde652798f, please check neutron logs for more information. [ 540.492398] env[62914]: ERROR nova.compute.manager [ 540.492398] env[62914]: Traceback (most recent call last): [ 540.492398] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 540.492398] env[62914]: listener.cb(fileno) [ 540.492398] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.492398] env[62914]: result = function(*args, **kwargs) [ 540.492398] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 540.492398] env[62914]: return func(*args, **kwargs) [ 540.492398] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.492398] env[62914]: raise e [ 540.492398] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.492398] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 540.492398] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.492398] env[62914]: created_port_ids = self._update_ports_for_instance( [ 540.492398] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.492398] env[62914]: with excutils.save_and_reraise_exception(): [ 540.492398] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.492398] env[62914]: self.force_reraise() [ 540.492398] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.492398] env[62914]: raise self.value [ 540.492398] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.492398] env[62914]: updated_port = self._update_port( [ 540.492398] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.492398] env[62914]: _ensure_no_port_binding_failure(port) [ 540.492398] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.492398] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 540.493344] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 0451d3e7-8933-4ae7-ad82-9ffde652798f, please check neutron logs for more information. [ 540.493344] env[62914]: Removing descriptor: 15 [ 540.493344] env[62914]: ERROR nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0451d3e7-8933-4ae7-ad82-9ffde652798f, please check neutron logs for more information. [ 540.493344] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Traceback (most recent call last): [ 540.493344] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 540.493344] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] yield resources [ 540.493344] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 540.493344] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] self.driver.spawn(context, instance, image_meta, [ 540.493344] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 540.493344] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.493344] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.493344] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] vm_ref = self.build_virtual_machine(instance, [ 540.493676] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.493676] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.493676] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.493676] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] for vif in network_info: [ 540.493676] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.493676] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] return self._sync_wrapper(fn, *args, **kwargs) [ 540.493676] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.493676] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] self.wait() [ 540.493676] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.493676] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] self[:] = self._gt.wait() [ 540.493676] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.493676] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] return self._exit_event.wait() [ 540.493676] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.494495] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] result = hub.switch() [ 540.494495] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.494495] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] return self.greenlet.switch() [ 540.494495] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.494495] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] result = function(*args, **kwargs) [ 540.494495] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 540.494495] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] return func(*args, **kwargs) [ 540.494495] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.494495] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] raise e [ 540.494495] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.494495] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] nwinfo = self.network_api.allocate_for_instance( [ 540.494495] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.494495] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] created_port_ids = self._update_ports_for_instance( [ 540.495434] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.495434] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] with excutils.save_and_reraise_exception(): [ 540.495434] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.495434] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] self.force_reraise() [ 540.495434] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.495434] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] raise self.value [ 540.495434] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.495434] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] updated_port = self._update_port( [ 540.495434] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.495434] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] _ensure_no_port_binding_failure(port) [ 540.495434] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.495434] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] raise exception.PortBindingFailed(port_id=port['id']) [ 540.495788] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] nova.exception.PortBindingFailed: Binding failed for port 0451d3e7-8933-4ae7-ad82-9ffde652798f, please check neutron logs for more information. [ 540.495788] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] [ 540.495788] env[62914]: INFO nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Terminating instance [ 540.496809] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "refresh_cache-df6047b8-8e4b-48e2-ae54-f5fda0aa77da" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.496887] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "refresh_cache-df6047b8-8e4b-48e2-ae54-f5fda0aa77da" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.497081] env[62914]: DEBUG nova.network.neutron [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 540.535419] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c758e8-e45e-06a7-91fa-6942c15db9b5, 'name': SearchDatastore_Task, 'duration_secs': 0.037281} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.536544] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.537437] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 540.539636] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.539819] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.540123] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 540.540388] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-36ced8fe-cfc2-4133-93e4-24a3825a6f92 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.561309] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 540.561493] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 540.562541] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca8be7c7-dca6-48ad-b88d-965ef4d2a97a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.570077] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 540.570077] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5239fd32-d62d-471f-5681-bee1126bd0c8" [ 540.570077] env[62914]: _type = "Task" [ 540.570077] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.580184] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5239fd32-d62d-471f-5681-bee1126bd0c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.706891] env[62914]: DEBUG nova.network.neutron [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.786277] env[62914]: DEBUG nova.network.neutron [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Successfully created port: 2dd6fb02-f6a2-42e3-987f-e9be0b492068 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 540.892688] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95cdc2c-7c6e-4cd5-8712-b939b7c71721 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.903396] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0adc2a55-68fd-488e-a02d-c51707f8cee6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.936217] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f61b634e-bd9a-4b9a-9a0c-ba4167fcbdd0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.944740] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0dfb31-f175-4de7-817f-21f27703a9d3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.960715] env[62914]: DEBUG nova.compute.provider_tree [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.993534] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.052611] env[62914]: DEBUG nova.network.neutron [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.088346] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5239fd32-d62d-471f-5681-bee1126bd0c8, 'name': SearchDatastore_Task, 'duration_secs': 0.011847} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.088723] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4eee1677-beac-4b59-ba42-655c0446f003 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.094407] env[62914]: DEBUG nova.network.neutron [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.096569] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 541.096569] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d97d5f-b2c5-c7eb-8471-2242518369ea" [ 541.096569] env[62914]: _type = "Task" [ 541.096569] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.108187] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d97d5f-b2c5-c7eb-8471-2242518369ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.423198] env[62914]: DEBUG nova.network.neutron [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.464554] env[62914]: DEBUG nova.scheduler.client.report [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.600984] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Releasing lock "refresh_cache-3db8507b-52b8-4eeb-b801-56e817c3b728" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.600984] env[62914]: DEBUG nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 541.600984] env[62914]: DEBUG nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 541.600984] env[62914]: DEBUG nova.network.neutron [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 541.613095] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d97d5f-b2c5-c7eb-8471-2242518369ea, 'name': SearchDatastore_Task, 'duration_secs': 0.016403} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.613434] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.613908] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417/3786cd93-1a50-4997-8894-840afa1c8417.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 541.614251] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5186f7a2-f870-44b2-9895-cefe558a6bf6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.626475] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 541.626475] env[62914]: value = "task-1352341" [ 541.626475] env[62914]: _type = "Task" [ 541.626475] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.636933] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352341, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.661264] env[62914]: DEBUG nova.network.neutron [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.933201] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "refresh_cache-df6047b8-8e4b-48e2-ae54-f5fda0aa77da" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.933955] env[62914]: DEBUG nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 541.933955] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 541.935624] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-041558eb-2fcc-4c10-bb1c-7ffe4fb2723f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.948267] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2d62c1-bae1-4806-a832-9a301f088eab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.970877] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.830s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.971533] env[62914]: ERROR nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 44ec2ba3-6f10-4fff-8e78-9872e8b8c0a0, please check neutron logs for more information. [ 541.971533] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Traceback (most recent call last): [ 541.971533] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 541.971533] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] self.driver.spawn(context, instance, image_meta, [ 541.971533] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 541.971533] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] self._vmops.spawn(context, instance, image_meta, injected_files, [ 541.971533] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 541.971533] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] vm_ref = self.build_virtual_machine(instance, [ 541.971533] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 541.971533] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] vif_infos = vmwarevif.get_vif_info(self._session, [ 541.971533] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 541.971884] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] for vif in network_info: [ 541.971884] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 541.971884] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] return self._sync_wrapper(fn, *args, **kwargs) [ 541.971884] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 541.971884] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] self.wait() [ 541.971884] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 541.971884] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] self[:] = self._gt.wait() [ 541.971884] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 541.971884] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] return self._exit_event.wait() [ 541.971884] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 541.971884] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] result = hub.switch() [ 541.971884] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 541.971884] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] return self.greenlet.switch() [ 541.972239] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.972239] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] result = function(*args, **kwargs) [ 541.972239] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 541.972239] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] return func(*args, **kwargs) [ 541.972239] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.972239] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] raise e [ 541.972239] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.972239] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] nwinfo = self.network_api.allocate_for_instance( [ 541.972239] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.972239] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] created_port_ids = self._update_ports_for_instance( [ 541.972239] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.972239] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] with excutils.save_and_reraise_exception(): [ 541.972239] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.972575] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] self.force_reraise() [ 541.972575] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.972575] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] raise self.value [ 541.972575] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.972575] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] updated_port = self._update_port( [ 541.972575] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.972575] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] _ensure_no_port_binding_failure(port) [ 541.972575] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.972575] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] raise exception.PortBindingFailed(port_id=port['id']) [ 541.972575] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] nova.exception.PortBindingFailed: Binding failed for port 44ec2ba3-6f10-4fff-8e78-9872e8b8c0a0, please check neutron logs for more information. [ 541.972575] env[62914]: ERROR nova.compute.manager [instance: 33f4159e-4181-4345-a716-0dfc409cee78] [ 541.972877] env[62914]: DEBUG nova.compute.utils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Binding failed for port 44ec2ba3-6f10-4fff-8e78-9872e8b8c0a0, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 541.983280] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.739s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.985637] env[62914]: INFO nova.compute.claims [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 541.989596] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance df6047b8-8e4b-48e2-ae54-f5fda0aa77da could not be found. [ 541.989991] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 541.989991] env[62914]: INFO nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Took 0.06 seconds to destroy the instance on the hypervisor. [ 541.993574] env[62914]: DEBUG oslo.service.loopingcall [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.994080] env[62914]: DEBUG nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Build of instance 33f4159e-4181-4345-a716-0dfc409cee78 was re-scheduled: Binding failed for port 44ec2ba3-6f10-4fff-8e78-9872e8b8c0a0, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 541.994915] env[62914]: DEBUG nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 541.995250] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Acquiring lock "refresh_cache-33f4159e-4181-4345-a716-0dfc409cee78" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.995402] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Acquired lock "refresh_cache-33f4159e-4181-4345-a716-0dfc409cee78" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.995558] env[62914]: DEBUG nova.network.neutron [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 541.996665] env[62914]: DEBUG nova.compute.manager [-] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 541.996665] env[62914]: DEBUG nova.network.neutron [-] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 542.040205] env[62914]: DEBUG nova.network.neutron [-] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.143212] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352341, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.165785] env[62914]: DEBUG nova.network.neutron [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.541785] env[62914]: DEBUG nova.network.neutron [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.544895] env[62914]: DEBUG nova.network.neutron [-] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.642050] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352341, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.881237} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.642350] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417/3786cd93-1a50-4997-8894-840afa1c8417.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 542.642627] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 542.642827] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-74e4fbed-d14c-4422-9c86-64ae43c4ff0a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.652576] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 542.652576] env[62914]: value = "task-1352342" [ 542.652576] env[62914]: _type = "Task" [ 542.652576] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.664588] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352342, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.668274] env[62914]: INFO nova.compute.manager [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] [instance: 3db8507b-52b8-4eeb-b801-56e817c3b728] Took 1.07 seconds to deallocate network for instance. [ 542.796562] env[62914]: DEBUG nova.network.neutron [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.051083] env[62914]: INFO nova.compute.manager [-] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Took 1.05 seconds to deallocate network for instance. [ 543.056612] env[62914]: DEBUG nova.compute.claims [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 543.057185] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.070308] env[62914]: DEBUG nova.compute.manager [req-57dce6bd-c904-468e-92dc-61b18e623ddf req-d7898ef7-95d1-4332-b020-b873e5337c60 service nova] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Received event network-vif-deleted-b615410e-0168-495f-b353-68103e3c6134 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 543.168655] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352342, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084481} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.171528] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 543.172466] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2233aef-5337-449f-b9c7-b7aa12d8dec2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.212050] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417/3786cd93-1a50-4997-8894-840afa1c8417.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 543.212050] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9a64a91-c57b-4e25-a53c-64dc1069b899 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.235065] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 543.235065] env[62914]: value = "task-1352343" [ 543.235065] env[62914]: _type = "Task" [ 543.235065] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.248036] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352343, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.254098] env[62914]: DEBUG nova.compute.manager [req-7250ff2d-2897-4282-ab9e-a6b0a934d40d req-375b5048-9979-49f3-951e-7af100d9b323 service nova] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Received event network-changed-0451d3e7-8933-4ae7-ad82-9ffde652798f {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 543.254332] env[62914]: DEBUG nova.compute.manager [req-7250ff2d-2897-4282-ab9e-a6b0a934d40d req-375b5048-9979-49f3-951e-7af100d9b323 service nova] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Refreshing instance network info cache due to event network-changed-0451d3e7-8933-4ae7-ad82-9ffde652798f. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 543.254539] env[62914]: DEBUG oslo_concurrency.lockutils [req-7250ff2d-2897-4282-ab9e-a6b0a934d40d req-375b5048-9979-49f3-951e-7af100d9b323 service nova] Acquiring lock "refresh_cache-df6047b8-8e4b-48e2-ae54-f5fda0aa77da" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.254674] env[62914]: DEBUG oslo_concurrency.lockutils [req-7250ff2d-2897-4282-ab9e-a6b0a934d40d req-375b5048-9979-49f3-951e-7af100d9b323 service nova] Acquired lock "refresh_cache-df6047b8-8e4b-48e2-ae54-f5fda0aa77da" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.254825] env[62914]: DEBUG nova.network.neutron [req-7250ff2d-2897-4282-ab9e-a6b0a934d40d req-375b5048-9979-49f3-951e-7af100d9b323 service nova] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Refreshing network info cache for port 0451d3e7-8933-4ae7-ad82-9ffde652798f {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 543.301388] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Releasing lock "refresh_cache-33f4159e-4181-4345-a716-0dfc409cee78" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.301388] env[62914]: DEBUG nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 543.301388] env[62914]: DEBUG nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 543.301388] env[62914]: DEBUG nova.network.neutron [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 543.316568] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e97676-aa6b-4cf0-88f2-80915b804753 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.328761] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52fe6205-c86f-472f-8520-51bb530197d5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.338276] env[62914]: DEBUG nova.network.neutron [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.372063] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a7d9912-999c-43f5-8cd9-acefd2fed4e7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.379366] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2587ef0e-de91-4b89-b803-dcd5e7f37087 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.398446] env[62914]: DEBUG nova.compute.provider_tree [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.690927] env[62914]: ERROR nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2dd6fb02-f6a2-42e3-987f-e9be0b492068, please check neutron logs for more information. [ 543.690927] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 543.690927] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.690927] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 543.690927] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.690927] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 543.690927] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.690927] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 543.690927] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.690927] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 543.690927] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.690927] env[62914]: ERROR nova.compute.manager raise self.value [ 543.690927] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.690927] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 543.690927] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.690927] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 543.691466] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.691466] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 543.691466] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2dd6fb02-f6a2-42e3-987f-e9be0b492068, please check neutron logs for more information. [ 543.691466] env[62914]: ERROR nova.compute.manager [ 543.691466] env[62914]: Traceback (most recent call last): [ 543.691466] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 543.691466] env[62914]: listener.cb(fileno) [ 543.691466] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.691466] env[62914]: result = function(*args, **kwargs) [ 543.691466] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 543.691466] env[62914]: return func(*args, **kwargs) [ 543.691466] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.691466] env[62914]: raise e [ 543.691466] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.691466] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 543.691466] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.691466] env[62914]: created_port_ids = self._update_ports_for_instance( [ 543.691466] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.691466] env[62914]: with excutils.save_and_reraise_exception(): [ 543.691466] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.691466] env[62914]: self.force_reraise() [ 543.691466] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.691466] env[62914]: raise self.value [ 543.691466] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.691466] env[62914]: updated_port = self._update_port( [ 543.691466] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.691466] env[62914]: _ensure_no_port_binding_failure(port) [ 543.691466] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.691466] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 543.692272] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 2dd6fb02-f6a2-42e3-987f-e9be0b492068, please check neutron logs for more information. [ 543.692272] env[62914]: Removing descriptor: 19 [ 543.692272] env[62914]: ERROR nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2dd6fb02-f6a2-42e3-987f-e9be0b492068, please check neutron logs for more information. [ 543.692272] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Traceback (most recent call last): [ 543.692272] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 543.692272] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] yield resources [ 543.692272] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 543.692272] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] self.driver.spawn(context, instance, image_meta, [ 543.692272] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 543.692272] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.692272] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.692272] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] vm_ref = self.build_virtual_machine(instance, [ 543.692592] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.692592] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.692592] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.692592] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] for vif in network_info: [ 543.692592] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 543.692592] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] return self._sync_wrapper(fn, *args, **kwargs) [ 543.692592] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 543.692592] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] self.wait() [ 543.692592] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 543.692592] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] self[:] = self._gt.wait() [ 543.692592] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.692592] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] return self._exit_event.wait() [ 543.692592] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.692950] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] result = hub.switch() [ 543.692950] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.692950] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] return self.greenlet.switch() [ 543.692950] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.692950] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] result = function(*args, **kwargs) [ 543.692950] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 543.692950] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] return func(*args, **kwargs) [ 543.692950] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.692950] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] raise e [ 543.692950] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.692950] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] nwinfo = self.network_api.allocate_for_instance( [ 543.692950] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.692950] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] created_port_ids = self._update_ports_for_instance( [ 543.693292] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.693292] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] with excutils.save_and_reraise_exception(): [ 543.693292] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.693292] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] self.force_reraise() [ 543.693292] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.693292] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] raise self.value [ 543.693292] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.693292] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] updated_port = self._update_port( [ 543.693292] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.693292] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] _ensure_no_port_binding_failure(port) [ 543.693292] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.693292] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] raise exception.PortBindingFailed(port_id=port['id']) [ 543.693642] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] nova.exception.PortBindingFailed: Binding failed for port 2dd6fb02-f6a2-42e3-987f-e9be0b492068, please check neutron logs for more information. [ 543.693642] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] [ 543.693642] env[62914]: INFO nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Terminating instance [ 543.697445] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Acquiring lock "refresh_cache-16b4e8bd-d84f-45a6-b531-3200ea545931" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.698104] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Acquired lock "refresh_cache-16b4e8bd-d84f-45a6-b531-3200ea545931" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.698104] env[62914]: DEBUG nova.network.neutron [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 543.715187] env[62914]: INFO nova.scheduler.client.report [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Deleted allocations for instance 3db8507b-52b8-4eeb-b801-56e817c3b728 [ 543.750303] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352343, 'name': ReconfigVM_Task, 'duration_secs': 0.328322} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.753788] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417/3786cd93-1a50-4997-8894-840afa1c8417.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 543.753788] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2b1da10d-21ff-43c3-976b-cc1436974900 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.767026] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 543.767026] env[62914]: value = "task-1352344" [ 543.767026] env[62914]: _type = "Task" [ 543.767026] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.781803] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352344, 'name': Rename_Task} progress is 10%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.812978] env[62914]: DEBUG nova.network.neutron [req-7250ff2d-2897-4282-ab9e-a6b0a934d40d req-375b5048-9979-49f3-951e-7af100d9b323 service nova] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.873499] env[62914]: DEBUG nova.network.neutron [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.905077] env[62914]: DEBUG nova.scheduler.client.report [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.988301] env[62914]: DEBUG nova.network.neutron [req-7250ff2d-2897-4282-ab9e-a6b0a934d40d req-375b5048-9979-49f3-951e-7af100d9b323 service nova] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.228315] env[62914]: DEBUG oslo_concurrency.lockutils [None req-04c25fe0-6555-45c4-a627-66d65e74c106 tempest-ImagesNegativeTestJSON-214482843 tempest-ImagesNegativeTestJSON-214482843-project-member] Lock "3db8507b-52b8-4eeb-b801-56e817c3b728" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.906s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.229624] env[62914]: DEBUG nova.network.neutron [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.274608] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352344, 'name': Rename_Task, 'duration_secs': 0.292761} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.274861] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 544.275285] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9260b805-ac45-40a4-aa6e-7ac26ff81c4b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.284606] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 544.284606] env[62914]: value = "task-1352345" [ 544.284606] env[62914]: _type = "Task" [ 544.284606] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.303102] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352345, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.309142] env[62914]: DEBUG nova.network.neutron [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.377368] env[62914]: INFO nova.compute.manager [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] [instance: 33f4159e-4181-4345-a716-0dfc409cee78] Took 1.08 seconds to deallocate network for instance. [ 544.417030] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.433s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.418196] env[62914]: DEBUG nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 544.421527] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.864s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.425545] env[62914]: INFO nova.compute.claims [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 544.492659] env[62914]: DEBUG oslo_concurrency.lockutils [req-7250ff2d-2897-4282-ab9e-a6b0a934d40d req-375b5048-9979-49f3-951e-7af100d9b323 service nova] Releasing lock "refresh_cache-df6047b8-8e4b-48e2-ae54-f5fda0aa77da" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.492954] env[62914]: DEBUG nova.compute.manager [req-7250ff2d-2897-4282-ab9e-a6b0a934d40d req-375b5048-9979-49f3-951e-7af100d9b323 service nova] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Received event network-vif-deleted-0451d3e7-8933-4ae7-ad82-9ffde652798f {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.732607] env[62914]: DEBUG nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 544.800722] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352345, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.813164] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Releasing lock "refresh_cache-16b4e8bd-d84f-45a6-b531-3200ea545931" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.813458] env[62914]: DEBUG nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 544.813719] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 544.814016] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1bb4ab71-53e6-4914-bbe4-7c3132975f1d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.829327] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396d8e66-8e9b-48d0-a615-3b311fa07679 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.859316] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 16b4e8bd-d84f-45a6-b531-3200ea545931 could not be found. [ 544.859316] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 544.859744] env[62914]: INFO nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Took 0.05 seconds to destroy the instance on the hypervisor. [ 544.860195] env[62914]: DEBUG oslo.service.loopingcall [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 544.860695] env[62914]: DEBUG nova.compute.manager [-] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 544.862934] env[62914]: DEBUG nova.network.neutron [-] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 544.891935] env[62914]: DEBUG nova.network.neutron [-] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.933421] env[62914]: DEBUG nova.compute.utils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 544.937523] env[62914]: DEBUG nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 544.937909] env[62914]: DEBUG nova.network.neutron [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 545.019736] env[62914]: DEBUG nova.policy [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd939241d75294f38ba67fc1bb9cb59cc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '230784fa3f0f4b9c980e2cef5cb39ef8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 545.256872] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.300317] env[62914]: DEBUG oslo_vmware.api [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352345, 'name': PowerOnVM_Task, 'duration_secs': 0.701224} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.300616] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 545.300824] env[62914]: DEBUG nova.compute.manager [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 545.301765] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0a81ff-8b4e-4e3f-ae2f-850e7f74c29e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.399959] env[62914]: DEBUG nova.network.neutron [-] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.435227] env[62914]: INFO nova.scheduler.client.report [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Deleted allocations for instance 33f4159e-4181-4345-a716-0dfc409cee78 [ 545.444307] env[62914]: DEBUG nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 545.534218] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "34b0b81e-fc98-4019-92c3-fe2ce2fdfd67" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.534218] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "34b0b81e-fc98-4019-92c3-fe2ce2fdfd67" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.623393] env[62914]: DEBUG nova.network.neutron [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Successfully created port: fe1003d5-35d9-4496-87f7-dcbf94483f17 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 545.680850] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304c4ee2-7d30-45db-9cec-d8ff92fe32c5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.690560] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81a2a59-6808-4eec-8457-ad503e68e835 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.723273] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f364ae-55a4-423b-bd23-71b0b8ab7d98 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.732765] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e382cb-03bf-47f5-9185-a61c06169b8c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.749142] env[62914]: DEBUG nova.compute.provider_tree [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.819153] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.904867] env[62914]: INFO nova.compute.manager [-] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Took 1.04 seconds to deallocate network for instance. [ 545.908548] env[62914]: DEBUG nova.compute.claims [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 545.908843] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.959583] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0cf5e0a8-a38d-40d6-918b-92fdc9802874 tempest-ImagesOneServerNegativeTestJSON-1651848507 tempest-ImagesOneServerNegativeTestJSON-1651848507-project-member] Lock "33f4159e-4181-4345-a716-0dfc409cee78" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.181s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.101326] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Acquiring lock "723105bc-6810-4759-8240-af0a128475e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.101571] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Lock "723105bc-6810-4759-8240-af0a128475e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.252516] env[62914]: DEBUG nova.scheduler.client.report [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.463220] env[62914]: DEBUG nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 546.465800] env[62914]: DEBUG nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 546.512554] env[62914]: DEBUG nova.virt.hardware [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 546.512844] env[62914]: DEBUG nova.virt.hardware [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 546.512942] env[62914]: DEBUG nova.virt.hardware [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 546.513217] env[62914]: DEBUG nova.virt.hardware [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 546.513387] env[62914]: DEBUG nova.virt.hardware [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 546.513570] env[62914]: DEBUG nova.virt.hardware [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 546.513802] env[62914]: DEBUG nova.virt.hardware [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 546.513967] env[62914]: DEBUG nova.virt.hardware [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 546.518073] env[62914]: DEBUG nova.virt.hardware [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 546.518258] env[62914]: DEBUG nova.virt.hardware [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 546.518432] env[62914]: DEBUG nova.virt.hardware [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 546.519897] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39dfdbe-6904-4ed1-9ce9-c7864e7a57c6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.534504] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7490417-8017-4dda-928d-0ce5f7f35769 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.658891] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Acquiring lock "bfcba422-5ed0-4bfd-83e9-cdde324d899f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.659268] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Lock "bfcba422-5ed0-4bfd-83e9-cdde324d899f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.760205] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.337s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.761646] env[62914]: DEBUG nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 546.767699] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.805s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.997861] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.195537] env[62914]: ERROR nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fe1003d5-35d9-4496-87f7-dcbf94483f17, please check neutron logs for more information. [ 547.195537] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 547.195537] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.195537] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 547.195537] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.195537] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 547.195537] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.195537] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 547.195537] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.195537] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 547.195537] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.195537] env[62914]: ERROR nova.compute.manager raise self.value [ 547.195537] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.195537] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 547.195537] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.195537] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 547.196362] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.196362] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 547.196362] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fe1003d5-35d9-4496-87f7-dcbf94483f17, please check neutron logs for more information. [ 547.196362] env[62914]: ERROR nova.compute.manager [ 547.196362] env[62914]: Traceback (most recent call last): [ 547.196362] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 547.196362] env[62914]: listener.cb(fileno) [ 547.196362] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.196362] env[62914]: result = function(*args, **kwargs) [ 547.196362] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 547.196362] env[62914]: return func(*args, **kwargs) [ 547.196362] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.196362] env[62914]: raise e [ 547.196362] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.196362] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 547.196362] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.196362] env[62914]: created_port_ids = self._update_ports_for_instance( [ 547.196362] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.196362] env[62914]: with excutils.save_and_reraise_exception(): [ 547.196362] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.196362] env[62914]: self.force_reraise() [ 547.196362] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.196362] env[62914]: raise self.value [ 547.196362] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.196362] env[62914]: updated_port = self._update_port( [ 547.196362] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.196362] env[62914]: _ensure_no_port_binding_failure(port) [ 547.196362] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.196362] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 547.197571] env[62914]: nova.exception.PortBindingFailed: Binding failed for port fe1003d5-35d9-4496-87f7-dcbf94483f17, please check neutron logs for more information. [ 547.197571] env[62914]: Removing descriptor: 19 [ 547.197571] env[62914]: ERROR nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fe1003d5-35d9-4496-87f7-dcbf94483f17, please check neutron logs for more information. [ 547.197571] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] Traceback (most recent call last): [ 547.197571] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 547.197571] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] yield resources [ 547.197571] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 547.197571] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] self.driver.spawn(context, instance, image_meta, [ 547.197571] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 547.197571] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.197571] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.197571] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] vm_ref = self.build_virtual_machine(instance, [ 547.198105] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.198105] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.198105] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.198105] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] for vif in network_info: [ 547.198105] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.198105] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] return self._sync_wrapper(fn, *args, **kwargs) [ 547.198105] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.198105] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] self.wait() [ 547.198105] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.198105] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] self[:] = self._gt.wait() [ 547.198105] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.198105] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] return self._exit_event.wait() [ 547.198105] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.198661] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] result = hub.switch() [ 547.198661] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.198661] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] return self.greenlet.switch() [ 547.198661] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.198661] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] result = function(*args, **kwargs) [ 547.198661] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 547.198661] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] return func(*args, **kwargs) [ 547.198661] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.198661] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] raise e [ 547.198661] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.198661] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] nwinfo = self.network_api.allocate_for_instance( [ 547.198661] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.198661] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] created_port_ids = self._update_ports_for_instance( [ 547.199633] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.199633] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] with excutils.save_and_reraise_exception(): [ 547.199633] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.199633] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] self.force_reraise() [ 547.199633] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.199633] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] raise self.value [ 547.199633] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.199633] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] updated_port = self._update_port( [ 547.199633] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.199633] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] _ensure_no_port_binding_failure(port) [ 547.199633] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.199633] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] raise exception.PortBindingFailed(port_id=port['id']) [ 547.200177] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] nova.exception.PortBindingFailed: Binding failed for port fe1003d5-35d9-4496-87f7-dcbf94483f17, please check neutron logs for more information. [ 547.200177] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] [ 547.200177] env[62914]: INFO nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Terminating instance [ 547.200177] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Acquiring lock "refresh_cache-a8761350-3044-4822-8221-ed108e21cd62" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.200177] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Acquired lock "refresh_cache-a8761350-3044-4822-8221-ed108e21cd62" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.200177] env[62914]: DEBUG nova.network.neutron [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 547.209998] env[62914]: DEBUG nova.compute.manager [req-4bf1532b-5d5c-4db9-a1c4-c0261115b179 req-451ac003-1c45-4bed-a371-18a704e3fe5e service nova] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Received event network-changed-2dd6fb02-f6a2-42e3-987f-e9be0b492068 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 547.210217] env[62914]: DEBUG nova.compute.manager [req-4bf1532b-5d5c-4db9-a1c4-c0261115b179 req-451ac003-1c45-4bed-a371-18a704e3fe5e service nova] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Refreshing instance network info cache due to event network-changed-2dd6fb02-f6a2-42e3-987f-e9be0b492068. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 547.210424] env[62914]: DEBUG oslo_concurrency.lockutils [req-4bf1532b-5d5c-4db9-a1c4-c0261115b179 req-451ac003-1c45-4bed-a371-18a704e3fe5e service nova] Acquiring lock "refresh_cache-16b4e8bd-d84f-45a6-b531-3200ea545931" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.210561] env[62914]: DEBUG oslo_concurrency.lockutils [req-4bf1532b-5d5c-4db9-a1c4-c0261115b179 req-451ac003-1c45-4bed-a371-18a704e3fe5e service nova] Acquired lock "refresh_cache-16b4e8bd-d84f-45a6-b531-3200ea545931" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.210713] env[62914]: DEBUG nova.network.neutron [req-4bf1532b-5d5c-4db9-a1c4-c0261115b179 req-451ac003-1c45-4bed-a371-18a704e3fe5e service nova] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Refreshing network info cache for port 2dd6fb02-f6a2-42e3-987f-e9be0b492068 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 547.269047] env[62914]: DEBUG nova.compute.utils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 547.270906] env[62914]: DEBUG nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 547.270994] env[62914]: DEBUG nova.network.neutron [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 547.329298] env[62914]: DEBUG nova.policy [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9fc6546d0209499c86b34b80e0ab1c50', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '57c073e6f15346bc8ca0d83b89bb167e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 547.596251] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bccd3fa-8033-4183-8734-7b911bd2a759 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.605136] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946c2ad9-63ed-49af-a16f-e68828b17190 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.646910] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d41a404-6382-4b1a-a3e2-08c3c9e9268a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.656423] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a26065b-0298-45c3-91df-8d43b9616361 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.673752] env[62914]: DEBUG nova.compute.provider_tree [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.727888] env[62914]: DEBUG nova.network.neutron [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.731883] env[62914]: DEBUG nova.network.neutron [req-4bf1532b-5d5c-4db9-a1c4-c0261115b179 req-451ac003-1c45-4bed-a371-18a704e3fe5e service nova] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.774537] env[62914]: DEBUG nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 547.805784] env[62914]: DEBUG nova.network.neutron [req-4bf1532b-5d5c-4db9-a1c4-c0261115b179 req-451ac003-1c45-4bed-a371-18a704e3fe5e service nova] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.828530] env[62914]: DEBUG nova.network.neutron [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.906693] env[62914]: DEBUG nova.network.neutron [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Successfully created port: 8f3888a9-3b63-4288-99ab-32523f8483a2 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.178282] env[62914]: DEBUG nova.scheduler.client.report [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 548.313602] env[62914]: DEBUG oslo_concurrency.lockutils [req-4bf1532b-5d5c-4db9-a1c4-c0261115b179 req-451ac003-1c45-4bed-a371-18a704e3fe5e service nova] Releasing lock "refresh_cache-16b4e8bd-d84f-45a6-b531-3200ea545931" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.314419] env[62914]: DEBUG nova.compute.manager [req-4bf1532b-5d5c-4db9-a1c4-c0261115b179 req-451ac003-1c45-4bed-a371-18a704e3fe5e service nova] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Received event network-vif-deleted-2dd6fb02-f6a2-42e3-987f-e9be0b492068 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 548.333724] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Releasing lock "refresh_cache-a8761350-3044-4822-8221-ed108e21cd62" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.334168] env[62914]: DEBUG nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 548.334376] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 548.334653] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-50065161-b7b9-41b2-9139-44aaa9e65d86 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.347701] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544674df-4f90-4713-9d73-3b003dbb9d10 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.379351] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a8761350-3044-4822-8221-ed108e21cd62 could not be found. [ 548.380023] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 548.380023] env[62914]: INFO nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Took 0.05 seconds to destroy the instance on the hypervisor. [ 548.380023] env[62914]: DEBUG oslo.service.loopingcall [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.380622] env[62914]: DEBUG nova.compute.manager [-] [instance: a8761350-3044-4822-8221-ed108e21cd62] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.381592] env[62914]: DEBUG nova.network.neutron [-] [instance: a8761350-3044-4822-8221-ed108e21cd62] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 548.414903] env[62914]: DEBUG nova.network.neutron [-] [instance: a8761350-3044-4822-8221-ed108e21cd62] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.683661] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.916s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.684326] env[62914]: ERROR nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 42db1776-19ce-4c94-b288-581cad28a7f9, please check neutron logs for more information. [ 548.684326] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Traceback (most recent call last): [ 548.684326] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 548.684326] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] self.driver.spawn(context, instance, image_meta, [ 548.684326] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 548.684326] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.684326] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.684326] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] vm_ref = self.build_virtual_machine(instance, [ 548.684326] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.684326] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.684326] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.684876] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] for vif in network_info: [ 548.684876] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 548.684876] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] return self._sync_wrapper(fn, *args, **kwargs) [ 548.684876] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 548.684876] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] self.wait() [ 548.684876] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 548.684876] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] self[:] = self._gt.wait() [ 548.684876] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.684876] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] return self._exit_event.wait() [ 548.684876] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.684876] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] result = hub.switch() [ 548.684876] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.684876] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] return self.greenlet.switch() [ 548.685307] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.685307] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] result = function(*args, **kwargs) [ 548.685307] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 548.685307] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] return func(*args, **kwargs) [ 548.685307] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.685307] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] raise e [ 548.685307] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.685307] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] nwinfo = self.network_api.allocate_for_instance( [ 548.685307] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 548.685307] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] created_port_ids = self._update_ports_for_instance( [ 548.685307] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 548.685307] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] with excutils.save_and_reraise_exception(): [ 548.685307] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.685721] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] self.force_reraise() [ 548.685721] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.685721] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] raise self.value [ 548.685721] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 548.685721] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] updated_port = self._update_port( [ 548.685721] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.685721] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] _ensure_no_port_binding_failure(port) [ 548.685721] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.685721] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] raise exception.PortBindingFailed(port_id=port['id']) [ 548.685721] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] nova.exception.PortBindingFailed: Binding failed for port 42db1776-19ce-4c94-b288-581cad28a7f9, please check neutron logs for more information. [ 548.685721] env[62914]: ERROR nova.compute.manager [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] [ 548.686059] env[62914]: DEBUG nova.compute.utils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Binding failed for port 42db1776-19ce-4c94-b288-581cad28a7f9, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 548.686726] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.082s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.688423] env[62914]: INFO nova.compute.claims [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 548.691569] env[62914]: DEBUG nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Build of instance a7ee1dbd-73d2-46be-bf43-008eebcd64e9 was re-scheduled: Binding failed for port 42db1776-19ce-4c94-b288-581cad28a7f9, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 548.692088] env[62914]: DEBUG nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 548.692315] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Acquiring lock "refresh_cache-a7ee1dbd-73d2-46be-bf43-008eebcd64e9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.692457] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Acquired lock "refresh_cache-a7ee1dbd-73d2-46be-bf43-008eebcd64e9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.692608] env[62914]: DEBUG nova.network.neutron [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 548.784499] env[62914]: DEBUG nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 548.812206] env[62914]: DEBUG nova.virt.hardware [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.812206] env[62914]: DEBUG nova.virt.hardware [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.812206] env[62914]: DEBUG nova.virt.hardware [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.812484] env[62914]: DEBUG nova.virt.hardware [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.812484] env[62914]: DEBUG nova.virt.hardware [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.812484] env[62914]: DEBUG nova.virt.hardware [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.812484] env[62914]: DEBUG nova.virt.hardware [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.812484] env[62914]: DEBUG nova.virt.hardware [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.812629] env[62914]: DEBUG nova.virt.hardware [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.812629] env[62914]: DEBUG nova.virt.hardware [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.812629] env[62914]: DEBUG nova.virt.hardware [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.813058] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fed1c49-4820-4493-80e5-fbdd2752eee4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.824068] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-727c0a26-50ed-416d-bc04-849891da6325 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.918415] env[62914]: DEBUG nova.network.neutron [-] [instance: a8761350-3044-4822-8221-ed108e21cd62] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.993551] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "f3509973-013e-45ce-87f9-357e782f26d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.993667] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "f3509973-013e-45ce-87f9-357e782f26d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.143902] env[62914]: ERROR nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8f3888a9-3b63-4288-99ab-32523f8483a2, please check neutron logs for more information. [ 549.143902] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 549.143902] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.143902] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 549.143902] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.143902] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 549.143902] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.143902] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 549.143902] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.143902] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 549.143902] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.143902] env[62914]: ERROR nova.compute.manager raise self.value [ 549.143902] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.143902] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 549.143902] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.143902] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 549.144909] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.144909] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 549.144909] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8f3888a9-3b63-4288-99ab-32523f8483a2, please check neutron logs for more information. [ 549.144909] env[62914]: ERROR nova.compute.manager [ 549.144909] env[62914]: Traceback (most recent call last): [ 549.144909] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 549.144909] env[62914]: listener.cb(fileno) [ 549.144909] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.144909] env[62914]: result = function(*args, **kwargs) [ 549.144909] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 549.144909] env[62914]: return func(*args, **kwargs) [ 549.144909] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.144909] env[62914]: raise e [ 549.144909] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.144909] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 549.144909] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.144909] env[62914]: created_port_ids = self._update_ports_for_instance( [ 549.144909] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.144909] env[62914]: with excutils.save_and_reraise_exception(): [ 549.144909] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.144909] env[62914]: self.force_reraise() [ 549.144909] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.144909] env[62914]: raise self.value [ 549.144909] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.144909] env[62914]: updated_port = self._update_port( [ 549.144909] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.144909] env[62914]: _ensure_no_port_binding_failure(port) [ 549.144909] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.144909] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 549.146965] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 8f3888a9-3b63-4288-99ab-32523f8483a2, please check neutron logs for more information. [ 549.146965] env[62914]: Removing descriptor: 19 [ 549.146965] env[62914]: ERROR nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8f3888a9-3b63-4288-99ab-32523f8483a2, please check neutron logs for more information. [ 549.146965] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Traceback (most recent call last): [ 549.146965] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 549.146965] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] yield resources [ 549.146965] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 549.146965] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] self.driver.spawn(context, instance, image_meta, [ 549.146965] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 549.146965] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.146965] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.146965] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] vm_ref = self.build_virtual_machine(instance, [ 549.147551] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.147551] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.147551] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.147551] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] for vif in network_info: [ 549.147551] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 549.147551] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] return self._sync_wrapper(fn, *args, **kwargs) [ 549.147551] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 549.147551] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] self.wait() [ 549.147551] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 549.147551] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] self[:] = self._gt.wait() [ 549.147551] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.147551] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] return self._exit_event.wait() [ 549.147551] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.147970] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] result = hub.switch() [ 549.147970] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.147970] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] return self.greenlet.switch() [ 549.147970] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.147970] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] result = function(*args, **kwargs) [ 549.147970] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 549.147970] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] return func(*args, **kwargs) [ 549.147970] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.147970] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] raise e [ 549.147970] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.147970] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] nwinfo = self.network_api.allocate_for_instance( [ 549.147970] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.147970] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] created_port_ids = self._update_ports_for_instance( [ 549.148995] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.148995] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] with excutils.save_and_reraise_exception(): [ 549.148995] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.148995] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] self.force_reraise() [ 549.148995] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.148995] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] raise self.value [ 549.148995] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.148995] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] updated_port = self._update_port( [ 549.148995] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.148995] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] _ensure_no_port_binding_failure(port) [ 549.148995] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.148995] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] raise exception.PortBindingFailed(port_id=port['id']) [ 549.149370] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] nova.exception.PortBindingFailed: Binding failed for port 8f3888a9-3b63-4288-99ab-32523f8483a2, please check neutron logs for more information. [ 549.149370] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] [ 549.149370] env[62914]: INFO nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Terminating instance [ 549.149370] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Acquiring lock "refresh_cache-856b3816-d4f8-4c9e-b74c-ea15d67976ef" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.149370] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Acquired lock "refresh_cache-856b3816-d4f8-4c9e-b74c-ea15d67976ef" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.149370] env[62914]: DEBUG nova.network.neutron [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 549.215944] env[62914]: DEBUG nova.network.neutron [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.365527] env[62914]: DEBUG nova.network.neutron [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.425065] env[62914]: INFO nova.compute.manager [-] [instance: a8761350-3044-4822-8221-ed108e21cd62] Took 1.04 seconds to deallocate network for instance. [ 549.426041] env[62914]: DEBUG nova.compute.claims [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 549.426486] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.513379] env[62914]: INFO nova.compute.manager [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Rebuilding instance [ 549.565494] env[62914]: DEBUG nova.compute.manager [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 549.566530] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a223225b-f383-416d-9b34-8152a0416819 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.686259] env[62914]: DEBUG nova.network.neutron [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.858921] env[62914]: DEBUG nova.network.neutron [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.872933] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Releasing lock "refresh_cache-a7ee1dbd-73d2-46be-bf43-008eebcd64e9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.873345] env[62914]: DEBUG nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 549.873643] env[62914]: DEBUG nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 549.873961] env[62914]: DEBUG nova.network.neutron [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 549.897138] env[62914]: DEBUG nova.network.neutron [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.950647] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c6b0bf-72c1-467b-8a00-baa978af520b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.960018] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e9f76f-85c7-4daf-a1ae-562232308ad5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.995214] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630928a7-6832-4dd9-9a4b-5440c3e8d885 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.004207] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f90b8c-ca69-4d60-8179-faac9d3031f8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.019114] env[62914]: DEBUG nova.compute.provider_tree [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.083680] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 550.083680] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-51b47ec8-16f9-4bad-b219-f117cb9e31ee {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.093390] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Waiting for the task: (returnval){ [ 550.093390] env[62914]: value = "task-1352346" [ 550.093390] env[62914]: _type = "Task" [ 550.093390] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.107906] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352346, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.308675] env[62914]: DEBUG nova.compute.manager [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] [instance: a8761350-3044-4822-8221-ed108e21cd62] Received event network-changed-fe1003d5-35d9-4496-87f7-dcbf94483f17 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 550.309124] env[62914]: DEBUG nova.compute.manager [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] [instance: a8761350-3044-4822-8221-ed108e21cd62] Refreshing instance network info cache due to event network-changed-fe1003d5-35d9-4496-87f7-dcbf94483f17. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 550.309339] env[62914]: DEBUG oslo_concurrency.lockutils [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] Acquiring lock "refresh_cache-a8761350-3044-4822-8221-ed108e21cd62" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.309568] env[62914]: DEBUG oslo_concurrency.lockutils [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] Acquired lock "refresh_cache-a8761350-3044-4822-8221-ed108e21cd62" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.310164] env[62914]: DEBUG nova.network.neutron [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] [instance: a8761350-3044-4822-8221-ed108e21cd62] Refreshing network info cache for port fe1003d5-35d9-4496-87f7-dcbf94483f17 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 550.365131] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Releasing lock "refresh_cache-856b3816-d4f8-4c9e-b74c-ea15d67976ef" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.365131] env[62914]: DEBUG nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 550.365131] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 550.365131] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a7f4e267-5690-4764-b211-c4406cc0def3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.376727] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701f0cba-a25f-4688-8aa2-fa0231738bd7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.402651] env[62914]: DEBUG nova.network.neutron [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.404941] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 856b3816-d4f8-4c9e-b74c-ea15d67976ef could not be found. [ 550.405418] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 550.405518] env[62914]: INFO nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Took 0.04 seconds to destroy the instance on the hypervisor. [ 550.406114] env[62914]: DEBUG oslo.service.loopingcall [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 550.406114] env[62914]: DEBUG nova.compute.manager [-] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 550.406692] env[62914]: DEBUG nova.network.neutron [-] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 550.429068] env[62914]: DEBUG nova.network.neutron [-] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.518581] env[62914]: DEBUG nova.compute.manager [None req-89c96056-3edd-4b7e-9cf1-08d52ea64bdf tempest-ServerDiagnosticsV248Test-2004122332 tempest-ServerDiagnosticsV248Test-2004122332-project-admin] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 550.520880] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577ad083-7e95-4c25-9883-537110f16b29 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.524521] env[62914]: DEBUG nova.scheduler.client.report [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 550.535506] env[62914]: INFO nova.compute.manager [None req-89c96056-3edd-4b7e-9cf1-08d52ea64bdf tempest-ServerDiagnosticsV248Test-2004122332 tempest-ServerDiagnosticsV248Test-2004122332-project-admin] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Retrieving diagnostics [ 550.536574] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b3fe6c-491c-425b-8672-5af5e0a388cb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.594815] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Acquiring lock "e11079b9-5f15-41d6-8c8a-c08dc7f63b2f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.595094] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Lock "e11079b9-5f15-41d6-8c8a-c08dc7f63b2f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.607887] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352346, 'name': PowerOffVM_Task, 'duration_secs': 0.183305} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.608176] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 550.608402] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 550.609312] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a712e76-e301-421e-af38-8620eab34903 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.617545] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 550.617799] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8a693c51-3c0b-4553-9e04-b831e55b0b5d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.649164] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 550.649508] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 550.649816] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Deleting the datastore file [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 550.650119] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-810562f5-d3e5-4bfa-be72-607f7de5a06b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.661376] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Waiting for the task: (returnval){ [ 550.661376] env[62914]: value = "task-1352348" [ 550.661376] env[62914]: _type = "Task" [ 550.661376] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.671701] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352348, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.833142] env[62914]: DEBUG nova.network.neutron [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] [instance: a8761350-3044-4822-8221-ed108e21cd62] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.905823] env[62914]: INFO nova.compute.manager [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] [instance: a7ee1dbd-73d2-46be-bf43-008eebcd64e9] Took 1.03 seconds to deallocate network for instance. [ 550.932456] env[62914]: DEBUG nova.network.neutron [-] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.945389] env[62914]: DEBUG nova.network.neutron [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] [instance: a8761350-3044-4822-8221-ed108e21cd62] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.030512] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.344s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.031036] env[62914]: DEBUG nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 551.033623] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.423s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.171396] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352348, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108289} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.171710] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 551.171922] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 551.172106] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 551.435597] env[62914]: INFO nova.compute.manager [-] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Took 1.03 seconds to deallocate network for instance. [ 551.442286] env[62914]: DEBUG nova.compute.claims [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 551.442668] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.452472] env[62914]: DEBUG oslo_concurrency.lockutils [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] Releasing lock "refresh_cache-a8761350-3044-4822-8221-ed108e21cd62" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.452749] env[62914]: DEBUG nova.compute.manager [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] [instance: a8761350-3044-4822-8221-ed108e21cd62] Received event network-vif-deleted-fe1003d5-35d9-4496-87f7-dcbf94483f17 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 551.452936] env[62914]: DEBUG nova.compute.manager [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Received event network-changed-8f3888a9-3b63-4288-99ab-32523f8483a2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 551.453190] env[62914]: DEBUG nova.compute.manager [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Refreshing instance network info cache due to event network-changed-8f3888a9-3b63-4288-99ab-32523f8483a2. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 551.453436] env[62914]: DEBUG oslo_concurrency.lockutils [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] Acquiring lock "refresh_cache-856b3816-d4f8-4c9e-b74c-ea15d67976ef" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.453575] env[62914]: DEBUG oslo_concurrency.lockutils [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] Acquired lock "refresh_cache-856b3816-d4f8-4c9e-b74c-ea15d67976ef" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.453731] env[62914]: DEBUG nova.network.neutron [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Refreshing network info cache for port 8f3888a9-3b63-4288-99ab-32523f8483a2 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 551.539099] env[62914]: DEBUG nova.compute.utils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 551.540514] env[62914]: DEBUG nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 551.540693] env[62914]: DEBUG nova.network.neutron [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 551.663627] env[62914]: DEBUG nova.policy [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c9da2a8d06e44e7b98cc5d76d34e0bb2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ace2b948d31443d195070fa73cfbe24d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 551.847553] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf14a376-60f6-4ec3-8c1e-3688a8123ac9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.856576] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731dbb3c-85b8-444d-8082-6c0ea6e9eff6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.890818] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a2d0e9-dfb7-4c6a-bae6-a2856d5dcf4d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.899726] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bdb637e-6eab-4c2b-a523-599ed5c1d09b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.916397] env[62914]: DEBUG nova.compute.provider_tree [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.957403] env[62914]: INFO nova.scheduler.client.report [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Deleted allocations for instance a7ee1dbd-73d2-46be-bf43-008eebcd64e9 [ 552.005922] env[62914]: DEBUG nova.network.neutron [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.044121] env[62914]: DEBUG nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 552.220829] env[62914]: DEBUG nova.virt.hardware [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 552.220829] env[62914]: DEBUG nova.virt.hardware [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 552.220829] env[62914]: DEBUG nova.virt.hardware [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 552.220829] env[62914]: DEBUG nova.virt.hardware [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 552.221113] env[62914]: DEBUG nova.virt.hardware [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 552.221113] env[62914]: DEBUG nova.virt.hardware [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 552.221113] env[62914]: DEBUG nova.virt.hardware [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 552.221113] env[62914]: DEBUG nova.virt.hardware [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 552.221878] env[62914]: DEBUG nova.virt.hardware [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 552.222378] env[62914]: DEBUG nova.virt.hardware [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 552.222595] env[62914]: DEBUG nova.virt.hardware [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 552.223566] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f65aba-bfd2-417b-bb1c-4161e47048cd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.232894] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38416cfc-8238-4bde-aeb0-e31b224c2560 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.248947] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Instance VIF info [] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 552.255331] env[62914]: DEBUG oslo.service.loopingcall [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 552.255638] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 552.255853] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f3743f87-8def-4270-bb00-d92259c4cce8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.281603] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 552.281603] env[62914]: value = "task-1352349" [ 552.281603] env[62914]: _type = "Task" [ 552.281603] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.290361] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352349, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.393472] env[62914]: DEBUG nova.network.neutron [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.419032] env[62914]: DEBUG nova.scheduler.client.report [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 552.470956] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7194890-8486-4e13-92c6-65edb7ae23fa tempest-AttachInterfacesUnderV243Test-1462606334 tempest-AttachInterfacesUnderV243Test-1462606334-project-member] Lock "a7ee1dbd-73d2-46be-bf43-008eebcd64e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.636s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.576669] env[62914]: DEBUG nova.network.neutron [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Successfully created port: 4301f15c-b22c-47fc-9616-26ee92aa1327 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 552.777493] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Acquiring lock "d5c90d13-41bc-4cca-abf3-5a5b13171a6f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.777493] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Lock "d5c90d13-41bc-4cca-abf3-5a5b13171a6f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.777493] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Acquiring lock "d5c90d13-41bc-4cca-abf3-5a5b13171a6f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.777493] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Lock "d5c90d13-41bc-4cca-abf3-5a5b13171a6f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.777712] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Lock "d5c90d13-41bc-4cca-abf3-5a5b13171a6f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.781331] env[62914]: INFO nova.compute.manager [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Terminating instance [ 552.786051] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Acquiring lock "refresh_cache-d5c90d13-41bc-4cca-abf3-5a5b13171a6f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.786281] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Acquired lock "refresh_cache-d5c90d13-41bc-4cca-abf3-5a5b13171a6f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.786494] env[62914]: DEBUG nova.network.neutron [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 552.794072] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352349, 'name': CreateVM_Task, 'duration_secs': 0.308512} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.794788] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 552.795519] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.795741] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.796120] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 552.796624] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-876c59d8-1e53-4ee3-ad1f-54c45bf59248 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.802896] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Waiting for the task: (returnval){ [ 552.802896] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52377288-eefc-a26a-68d9-a79d3c9fea30" [ 552.802896] env[62914]: _type = "Task" [ 552.802896] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.813477] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52377288-eefc-a26a-68d9-a79d3c9fea30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.897037] env[62914]: DEBUG oslo_concurrency.lockutils [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] Releasing lock "refresh_cache-856b3816-d4f8-4c9e-b74c-ea15d67976ef" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.897037] env[62914]: DEBUG nova.compute.manager [req-6e6cacd6-4f8a-4583-b503-e1fd5c44ecf5 req-eabf8069-46eb-4580-ac24-bc7886204d9d service nova] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Received event network-vif-deleted-8f3888a9-3b63-4288-99ab-32523f8483a2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 552.928430] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.895s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.929930] env[62914]: ERROR nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b615410e-0168-495f-b353-68103e3c6134, please check neutron logs for more information. [ 552.929930] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Traceback (most recent call last): [ 552.929930] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 552.929930] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] self.driver.spawn(context, instance, image_meta, [ 552.929930] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 552.929930] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] self._vmops.spawn(context, instance, image_meta, injected_files, [ 552.929930] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 552.929930] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] vm_ref = self.build_virtual_machine(instance, [ 552.929930] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 552.929930] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] vif_infos = vmwarevif.get_vif_info(self._session, [ 552.929930] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 552.930515] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] for vif in network_info: [ 552.930515] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 552.930515] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] return self._sync_wrapper(fn, *args, **kwargs) [ 552.930515] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 552.930515] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] self.wait() [ 552.930515] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 552.930515] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] self[:] = self._gt.wait() [ 552.930515] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 552.930515] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] return self._exit_event.wait() [ 552.930515] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 552.930515] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] result = hub.switch() [ 552.930515] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 552.930515] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] return self.greenlet.switch() [ 552.931583] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 552.931583] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] result = function(*args, **kwargs) [ 552.931583] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 552.931583] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] return func(*args, **kwargs) [ 552.931583] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 552.931583] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] raise e [ 552.931583] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.931583] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] nwinfo = self.network_api.allocate_for_instance( [ 552.931583] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 552.931583] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] created_port_ids = self._update_ports_for_instance( [ 552.931583] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 552.931583] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] with excutils.save_and_reraise_exception(): [ 552.931583] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.932322] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] self.force_reraise() [ 552.932322] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.932322] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] raise self.value [ 552.932322] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 552.932322] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] updated_port = self._update_port( [ 552.932322] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.932322] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] _ensure_no_port_binding_failure(port) [ 552.932322] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.932322] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] raise exception.PortBindingFailed(port_id=port['id']) [ 552.932322] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] nova.exception.PortBindingFailed: Binding failed for port b615410e-0168-495f-b353-68103e3c6134, please check neutron logs for more information. [ 552.932322] env[62914]: ERROR nova.compute.manager [instance: c78502a7-266d-49cf-9d50-cc89ed266647] [ 552.932612] env[62914]: DEBUG nova.compute.utils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Binding failed for port b615410e-0168-495f-b353-68103e3c6134, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 552.933702] env[62914]: DEBUG nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Build of instance c78502a7-266d-49cf-9d50-cc89ed266647 was re-scheduled: Binding failed for port b615410e-0168-495f-b353-68103e3c6134, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 552.934390] env[62914]: DEBUG nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 552.934614] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Acquiring lock "refresh_cache-c78502a7-266d-49cf-9d50-cc89ed266647" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.934762] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Acquired lock "refresh_cache-c78502a7-266d-49cf-9d50-cc89ed266647" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.934916] env[62914]: DEBUG nova.network.neutron [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 552.937244] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 12.689s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.976106] env[62914]: DEBUG nova.compute.manager [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 553.056805] env[62914]: DEBUG nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 553.091238] env[62914]: DEBUG nova.virt.hardware [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:09:03Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='915220673',id=25,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-2094245477',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 553.091533] env[62914]: DEBUG nova.virt.hardware [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 553.092118] env[62914]: DEBUG nova.virt.hardware [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 553.092118] env[62914]: DEBUG nova.virt.hardware [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 553.092118] env[62914]: DEBUG nova.virt.hardware [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 553.093492] env[62914]: DEBUG nova.virt.hardware [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 553.093725] env[62914]: DEBUG nova.virt.hardware [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 553.093885] env[62914]: DEBUG nova.virt.hardware [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 553.095045] env[62914]: DEBUG nova.virt.hardware [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 553.095864] env[62914]: DEBUG nova.virt.hardware [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 553.095864] env[62914]: DEBUG nova.virt.hardware [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 553.096519] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af63dc07-9b5e-4fae-b42e-1d8711b9eeff {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.111348] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d64374d7-098b-448a-8c9f-0d82ef42aeb0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.317571] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52377288-eefc-a26a-68d9-a79d3c9fea30, 'name': SearchDatastore_Task, 'duration_secs': 0.009741} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.317571] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.317571] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 553.317571] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.318061] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.318061] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 553.318061] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-203838e4-c8a8-4fe9-95c3-cb197a771167 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.328892] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 553.329153] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 553.330243] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-daa1379d-3951-4b74-83bd-15057a7b912b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.334183] env[62914]: DEBUG nova.network.neutron [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.340132] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Waiting for the task: (returnval){ [ 553.340132] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5213c3d3-e3f3-5f4f-d3a8-c8e543d0be8b" [ 553.340132] env[62914]: _type = "Task" [ 553.340132] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.353217] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5213c3d3-e3f3-5f4f-d3a8-c8e543d0be8b, 'name': SearchDatastore_Task, 'duration_secs': 0.009185} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.354415] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59f7e719-6830-412b-a44d-61751369e2fe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.360345] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Waiting for the task: (returnval){ [ 553.360345] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]527af987-be07-30ec-b3c3-42d0c52af812" [ 553.360345] env[62914]: _type = "Task" [ 553.360345] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.369890] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]527af987-be07-30ec-b3c3-42d0c52af812, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.494175] env[62914]: DEBUG nova.network.neutron [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.503363] env[62914]: DEBUG nova.network.neutron [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.514176] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.642352] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Acquiring lock "8213bc31-1bd5-40b6-99a0-1254b9517b6a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.642884] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Lock "8213bc31-1bd5-40b6-99a0-1254b9517b6a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.765014] env[62914]: DEBUG nova.network.neutron [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.876325] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]527af987-be07-30ec-b3c3-42d0c52af812, 'name': SearchDatastore_Task, 'duration_secs': 0.008882} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.876939] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.878564] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417/3786cd93-1a50-4997-8894-840afa1c8417.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 553.878564] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5c9285b-01d3-48f0-b879-3ea8ace34eb7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.888968] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Waiting for the task: (returnval){ [ 553.888968] env[62914]: value = "task-1352350" [ 553.888968] env[62914]: _type = "Task" [ 553.888968] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.902284] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352350, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.981460] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 3786cd93-1a50-4997-8894-840afa1c8417 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 553.981460] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance d5c90d13-41bc-4cca-abf3-5a5b13171a6f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 554.008559] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Releasing lock "refresh_cache-d5c90d13-41bc-4cca-abf3-5a5b13171a6f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.008559] env[62914]: DEBUG nova.compute.manager [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 554.008559] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 554.008559] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28250cbf-42ff-4d5c-8c30-043bcb6e38f4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.019497] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 554.019827] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0483246-d38a-40f3-83e6-9bbe3741d746 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.027408] env[62914]: DEBUG oslo_vmware.api [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for the task: (returnval){ [ 554.027408] env[62914]: value = "task-1352351" [ 554.027408] env[62914]: _type = "Task" [ 554.027408] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.040702] env[62914]: DEBUG oslo_vmware.api [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352351, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.268917] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Releasing lock "refresh_cache-c78502a7-266d-49cf-9d50-cc89ed266647" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.269170] env[62914]: DEBUG nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 554.269360] env[62914]: DEBUG nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 554.270156] env[62914]: DEBUG nova.network.neutron [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 554.331101] env[62914]: DEBUG nova.network.neutron [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.401894] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352350, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.484187] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance c78502a7-266d-49cf-9d50-cc89ed266647 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 554.484379] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance df6047b8-8e4b-48e2-ae54-f5fda0aa77da actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 554.486035] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 16b4e8bd-d84f-45a6-b531-3200ea545931 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 554.486035] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance a8761350-3044-4822-8221-ed108e21cd62 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 554.486035] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 856b3816-d4f8-4c9e-b74c-ea15d67976ef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 554.486035] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 889af877-29c4-4c06-9995-49fb15b21a55 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 554.537705] env[62914]: DEBUG oslo_vmware.api [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352351, 'name': PowerOffVM_Task, 'duration_secs': 0.152858} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.537967] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 554.540460] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 554.540460] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-00d9cd7d-9080-49dd-8990-3035fc2c739a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.566222] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 554.566459] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 554.566653] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Deleting the datastore file [datastore1] d5c90d13-41bc-4cca-abf3-5a5b13171a6f {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 554.566912] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12e98ba2-6fb0-49ec-bd77-632e092ce752 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.573750] env[62914]: DEBUG oslo_vmware.api [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for the task: (returnval){ [ 554.573750] env[62914]: value = "task-1352353" [ 554.573750] env[62914]: _type = "Task" [ 554.573750] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.584345] env[62914]: DEBUG oslo_vmware.api [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352353, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.834932] env[62914]: DEBUG nova.network.neutron [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.901291] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352350, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.536227} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.901545] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417/3786cd93-1a50-4997-8894-840afa1c8417.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 554.901760] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 554.902012] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-efebef4c-402b-47f3-8735-cabaeefec186 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.909210] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Waiting for the task: (returnval){ [ 554.909210] env[62914]: value = "task-1352354" [ 554.909210] env[62914]: _type = "Task" [ 554.909210] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.918869] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352354, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.988426] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 4981a04f-98e6-474b-850a-634d977d9e42 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.088643] env[62914]: DEBUG oslo_vmware.api [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Task: {'id': task-1352353, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093325} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.091440] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 555.091440] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 555.091440] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 555.091440] env[62914]: INFO nova.compute.manager [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Took 1.08 seconds to destroy the instance on the hypervisor. [ 555.091440] env[62914]: DEBUG oslo.service.loopingcall [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 555.091968] env[62914]: DEBUG nova.compute.manager [-] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 555.091968] env[62914]: DEBUG nova.network.neutron [-] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 555.116129] env[62914]: DEBUG nova.network.neutron [-] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.337797] env[62914]: INFO nova.compute.manager [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: c78502a7-266d-49cf-9d50-cc89ed266647] Took 1.07 seconds to deallocate network for instance. [ 555.427127] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352354, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067647} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.427127] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 555.427127] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd3c45a-da4e-43f7-9a98-026a18391932 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.455189] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417/3786cd93-1a50-4997-8894-840afa1c8417.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 555.455584] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ccbe953-1323-43b4-9041-e10a04d606f6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.482077] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Waiting for the task: (returnval){ [ 555.482077] env[62914]: value = "task-1352355" [ 555.482077] env[62914]: _type = "Task" [ 555.482077] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.492561] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352355, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.494418] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance f1cee1cb-aceb-47e8-b2da-b94563f922e2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.618449] env[62914]: DEBUG nova.network.neutron [-] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.992747] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352355, 'name': ReconfigVM_Task, 'duration_secs': 0.274172} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.993089] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417/3786cd93-1a50-4997-8894-840afa1c8417.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 555.993995] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-547b3448-21d5-45b1-ad24-f516e0bb25bd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.996528] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 6ec0c746-2bb8-4387-8b31-b49b0d708614 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 556.001947] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Waiting for the task: (returnval){ [ 556.001947] env[62914]: value = "task-1352356" [ 556.001947] env[62914]: _type = "Task" [ 556.001947] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.013261] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352356, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.121528] env[62914]: INFO nova.compute.manager [-] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Took 1.03 seconds to deallocate network for instance. [ 556.390370] env[62914]: INFO nova.scheduler.client.report [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Deleted allocations for instance c78502a7-266d-49cf-9d50-cc89ed266647 [ 556.445931] env[62914]: ERROR nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4301f15c-b22c-47fc-9616-26ee92aa1327, please check neutron logs for more information. [ 556.445931] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 556.445931] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.445931] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 556.445931] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.445931] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 556.445931] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.445931] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 556.445931] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.445931] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 556.445931] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.445931] env[62914]: ERROR nova.compute.manager raise self.value [ 556.445931] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.445931] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 556.445931] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.445931] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 556.446525] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.446525] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 556.446525] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4301f15c-b22c-47fc-9616-26ee92aa1327, please check neutron logs for more information. [ 556.446525] env[62914]: ERROR nova.compute.manager [ 556.446525] env[62914]: Traceback (most recent call last): [ 556.446525] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 556.446525] env[62914]: listener.cb(fileno) [ 556.446525] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.446525] env[62914]: result = function(*args, **kwargs) [ 556.446525] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 556.446525] env[62914]: return func(*args, **kwargs) [ 556.446525] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.446525] env[62914]: raise e [ 556.446525] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.446525] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 556.446525] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.446525] env[62914]: created_port_ids = self._update_ports_for_instance( [ 556.446525] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.446525] env[62914]: with excutils.save_and_reraise_exception(): [ 556.446525] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.446525] env[62914]: self.force_reraise() [ 556.446525] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.446525] env[62914]: raise self.value [ 556.446525] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.446525] env[62914]: updated_port = self._update_port( [ 556.446525] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.446525] env[62914]: _ensure_no_port_binding_failure(port) [ 556.446525] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.446525] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 556.447384] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 4301f15c-b22c-47fc-9616-26ee92aa1327, please check neutron logs for more information. [ 556.447384] env[62914]: Removing descriptor: 19 [ 556.447384] env[62914]: ERROR nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4301f15c-b22c-47fc-9616-26ee92aa1327, please check neutron logs for more information. [ 556.447384] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Traceback (most recent call last): [ 556.447384] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 556.447384] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] yield resources [ 556.447384] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.447384] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] self.driver.spawn(context, instance, image_meta, [ 556.447384] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 556.447384] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.447384] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.447384] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] vm_ref = self.build_virtual_machine(instance, [ 556.447801] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.447801] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.447801] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.447801] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] for vif in network_info: [ 556.447801] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.447801] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] return self._sync_wrapper(fn, *args, **kwargs) [ 556.447801] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.447801] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] self.wait() [ 556.447801] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.447801] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] self[:] = self._gt.wait() [ 556.447801] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.447801] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] return self._exit_event.wait() [ 556.447801] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.448217] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] result = hub.switch() [ 556.448217] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.448217] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] return self.greenlet.switch() [ 556.448217] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.448217] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] result = function(*args, **kwargs) [ 556.448217] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 556.448217] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] return func(*args, **kwargs) [ 556.448217] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.448217] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] raise e [ 556.448217] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.448217] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] nwinfo = self.network_api.allocate_for_instance( [ 556.448217] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.448217] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] created_port_ids = self._update_ports_for_instance( [ 556.448665] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.448665] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] with excutils.save_and_reraise_exception(): [ 556.448665] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.448665] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] self.force_reraise() [ 556.448665] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.448665] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] raise self.value [ 556.448665] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.448665] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] updated_port = self._update_port( [ 556.448665] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.448665] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] _ensure_no_port_binding_failure(port) [ 556.448665] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.448665] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] raise exception.PortBindingFailed(port_id=port['id']) [ 556.453690] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] nova.exception.PortBindingFailed: Binding failed for port 4301f15c-b22c-47fc-9616-26ee92aa1327, please check neutron logs for more information. [ 556.453690] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] [ 556.453690] env[62914]: INFO nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Terminating instance [ 556.453690] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Acquiring lock "refresh_cache-889af877-29c4-4c06-9995-49fb15b21a55" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.453690] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Acquired lock "refresh_cache-889af877-29c4-4c06-9995-49fb15b21a55" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.453690] env[62914]: DEBUG nova.network.neutron [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 556.500097] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 556.519563] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352356, 'name': Rename_Task, 'duration_secs': 0.286899} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.519841] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 556.521651] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4048b3d6-ec85-4b22-b0dc-a5970b83232f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.528930] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Waiting for the task: (returnval){ [ 556.528930] env[62914]: value = "task-1352357" [ 556.528930] env[62914]: _type = "Task" [ 556.528930] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.538415] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352357, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.631908] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.906135] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80c744f6-bc9d-4d2a-9374-4435fac1a91e tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Lock "c78502a7-266d-49cf-9d50-cc89ed266647" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.569s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.000996] env[62914]: DEBUG nova.network.neutron [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.007905] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 723105bc-6810-4759-8240-af0a128475e4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 557.042355] env[62914]: DEBUG oslo_vmware.api [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Task: {'id': task-1352357, 'name': PowerOnVM_Task, 'duration_secs': 0.407627} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.042355] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 557.042471] env[62914]: DEBUG nova.compute.manager [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 557.043220] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2eb83ab-c4e0-480a-8288-d19465b67123 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.119113] env[62914]: DEBUG nova.compute.manager [req-9cd36589-27af-4d72-adb8-b063d5c6eec1 req-733f7ca4-7a25-4eed-9803-5d68770f1875 service nova] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Received event network-changed-4301f15c-b22c-47fc-9616-26ee92aa1327 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 557.119352] env[62914]: DEBUG nova.compute.manager [req-9cd36589-27af-4d72-adb8-b063d5c6eec1 req-733f7ca4-7a25-4eed-9803-5d68770f1875 service nova] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Refreshing instance network info cache due to event network-changed-4301f15c-b22c-47fc-9616-26ee92aa1327. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 557.119485] env[62914]: DEBUG oslo_concurrency.lockutils [req-9cd36589-27af-4d72-adb8-b063d5c6eec1 req-733f7ca4-7a25-4eed-9803-5d68770f1875 service nova] Acquiring lock "refresh_cache-889af877-29c4-4c06-9995-49fb15b21a55" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.409504] env[62914]: DEBUG nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 557.465162] env[62914]: DEBUG nova.network.neutron [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.512626] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance bfcba422-5ed0-4bfd-83e9-cdde324d899f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 557.566155] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.946459] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.970026] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Releasing lock "refresh_cache-889af877-29c4-4c06-9995-49fb15b21a55" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.970468] env[62914]: DEBUG nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 557.971473] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 557.971473] env[62914]: DEBUG oslo_concurrency.lockutils [req-9cd36589-27af-4d72-adb8-b063d5c6eec1 req-733f7ca4-7a25-4eed-9803-5d68770f1875 service nova] Acquired lock "refresh_cache-889af877-29c4-4c06-9995-49fb15b21a55" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.971473] env[62914]: DEBUG nova.network.neutron [req-9cd36589-27af-4d72-adb8-b063d5c6eec1 req-733f7ca4-7a25-4eed-9803-5d68770f1875 service nova] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Refreshing network info cache for port 4301f15c-b22c-47fc-9616-26ee92aa1327 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 557.972372] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fcbe3e05-4854-4465-879b-1be7a374fd95 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.989191] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00053f27-c01e-4805-a42e-110cc4529559 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.017678] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 889af877-29c4-4c06-9995-49fb15b21a55 could not be found. [ 558.017933] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 558.018296] env[62914]: INFO nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Took 0.05 seconds to destroy the instance on the hypervisor. [ 558.018590] env[62914]: DEBUG oslo.service.loopingcall [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 558.019283] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance f3509973-013e-45ce-87f9-357e782f26d7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 558.021136] env[62914]: DEBUG nova.compute.manager [-] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 558.021286] env[62914]: DEBUG nova.network.neutron [-] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 558.088530] env[62914]: DEBUG nova.network.neutron [-] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.506307] env[62914]: DEBUG nova.network.neutron [req-9cd36589-27af-4d72-adb8-b063d5c6eec1 req-733f7ca4-7a25-4eed-9803-5d68770f1875 service nova] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.522423] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance e11079b9-5f15-41d6-8c8a-c08dc7f63b2f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 558.593149] env[62914]: DEBUG nova.network.neutron [-] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.792152] env[62914]: DEBUG nova.network.neutron [req-9cd36589-27af-4d72-adb8-b063d5c6eec1 req-733f7ca4-7a25-4eed-9803-5d68770f1875 service nova] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.027048] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 8213bc31-1bd5-40b6-99a0-1254b9517b6a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 559.027334] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 559.027541] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 559.101830] env[62914]: INFO nova.compute.manager [-] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Took 1.08 seconds to deallocate network for instance. [ 559.104768] env[62914]: DEBUG nova.compute.claims [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 559.105341] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.295441] env[62914]: DEBUG oslo_concurrency.lockutils [req-9cd36589-27af-4d72-adb8-b063d5c6eec1 req-733f7ca4-7a25-4eed-9803-5d68770f1875 service nova] Releasing lock "refresh_cache-889af877-29c4-4c06-9995-49fb15b21a55" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.337174] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2369ca4a-cd4a-4d9c-b5b2-c3c5eb56b4b0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.345379] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4596bf3c-151b-4380-9b82-143ed6421975 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.379306] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c31e91-f439-41cc-b030-83670e340377 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.387646] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c49189-d3c8-4b3a-afc6-655b7ce3136b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.405058] env[62914]: DEBUG nova.compute.provider_tree [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.908529] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 560.274499] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Acquiring lock "5e695914-afa5-4f79-972d-852a4ac8aa20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.274945] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Lock "5e695914-afa5-4f79-972d-852a4ac8aa20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.329183] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquiring lock "3786cd93-1a50-4997-8894-840afa1c8417" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.329332] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Lock "3786cd93-1a50-4997-8894-840afa1c8417" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.329534] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquiring lock "3786cd93-1a50-4997-8894-840afa1c8417-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.329703] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Lock "3786cd93-1a50-4997-8894-840afa1c8417-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.329864] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Lock "3786cd93-1a50-4997-8894-840afa1c8417-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.334715] env[62914]: INFO nova.compute.manager [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Terminating instance [ 560.339569] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquiring lock "refresh_cache-3786cd93-1a50-4997-8894-840afa1c8417" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.339731] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquired lock "refresh_cache-3786cd93-1a50-4997-8894-840afa1c8417" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.339951] env[62914]: DEBUG nova.network.neutron [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 560.418511] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 560.418511] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.479s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.418511] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.423s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.418511] env[62914]: INFO nova.compute.claims [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 560.472748] env[62914]: DEBUG nova.compute.manager [req-985e518b-76ed-4afd-8374-ac1d33d80ad5 req-ea5f6640-6bad-463e-b9be-bc844152b916 service nova] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Received event network-vif-deleted-4301f15c-b22c-47fc-9616-26ee92aa1327 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.875473] env[62914]: DEBUG nova.network.neutron [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.041169] env[62914]: DEBUG nova.network.neutron [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.546378] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Releasing lock "refresh_cache-3786cd93-1a50-4997-8894-840afa1c8417" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.546661] env[62914]: DEBUG nova.compute.manager [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 561.546703] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 561.547664] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36312df5-7db5-4142-af21-29e816bdece1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.559327] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 561.559650] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd964b9d-25f4-4a73-b98c-4e967ef91754 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.569530] env[62914]: DEBUG oslo_vmware.api [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 561.569530] env[62914]: value = "task-1352358" [ 561.569530] env[62914]: _type = "Task" [ 561.569530] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.586642] env[62914]: DEBUG oslo_vmware.api [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352358, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.757889] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac56610-9dd2-4e6b-9df5-e62fbeaf1612 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.767475] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1319c46f-e1ae-4d29-8f3e-8bdcac428110 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.802339] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb9af8a-6ce9-4ca7-9038-233f48c79244 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.811168] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23842d62-3457-4a97-825f-4d59983ecd5f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.826457] env[62914]: DEBUG nova.compute.provider_tree [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.085496] env[62914]: DEBUG oslo_vmware.api [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352358, 'name': PowerOffVM_Task, 'duration_secs': 0.141425} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.085762] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 562.086129] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 562.086361] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0a0736c9-7013-4cc6-982a-a8ec29c6f73f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.113433] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 562.113654] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 562.114154] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Deleting the datastore file [datastore2] 3786cd93-1a50-4997-8894-840afa1c8417 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 562.114154] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a7cc5ebb-5573-4238-9cfc-094178bd8a82 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.126019] env[62914]: DEBUG oslo_vmware.api [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for the task: (returnval){ [ 562.126019] env[62914]: value = "task-1352360" [ 562.126019] env[62914]: _type = "Task" [ 562.126019] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.134771] env[62914]: DEBUG oslo_vmware.api [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352360, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.164273] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Acquiring lock "964f98e4-3462-4362-b551-576302776f3a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.164528] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Lock "964f98e4-3462-4362-b551-576302776f3a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.329629] env[62914]: DEBUG nova.scheduler.client.report [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 562.639886] env[62914]: DEBUG oslo_vmware.api [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Task: {'id': task-1352360, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133267} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.640784] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 562.640784] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 562.640784] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 562.640784] env[62914]: INFO nova.compute.manager [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Took 1.09 seconds to destroy the instance on the hypervisor. [ 562.641147] env[62914]: DEBUG oslo.service.loopingcall [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 562.641555] env[62914]: DEBUG nova.compute.manager [-] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.641555] env[62914]: DEBUG nova.network.neutron [-] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 562.695269] env[62914]: DEBUG nova.network.neutron [-] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.835679] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.835679] env[62914]: DEBUG nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 562.840517] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.784s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.202020] env[62914]: DEBUG nova.network.neutron [-] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.339053] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "283ffd74-3891-4b5e-bcf5-49b265978051" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.339317] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "283ffd74-3891-4b5e-bcf5-49b265978051" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.353052] env[62914]: DEBUG nova.compute.utils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 563.354578] env[62914]: DEBUG nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 563.355349] env[62914]: DEBUG nova.network.neutron [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 563.435391] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "bfcba5cb-d9c4-4ddd-9018-d316c13525d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.435391] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "bfcba5cb-d9c4-4ddd-9018-d316c13525d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.004s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.490018] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "86784639-8ac7-4f67-be5b-05c3ab1229b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.490528] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "86784639-8ac7-4f67-be5b-05c3ab1229b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.506175] env[62914]: DEBUG nova.policy [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1b78520da3e4766b11a2bb58f7835b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c08bc55c33c3464eae078eaef8b0f8bf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 563.702118] env[62914]: INFO nova.compute.manager [-] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Took 1.06 seconds to deallocate network for instance. [ 563.783989] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34e3500-37cf-4ad2-a434-040adea65be5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.793373] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df58568-1725-4af8-ae9b-197a767cded3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.839076] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35e128b8-f7e3-4358-8070-7f837a4acecf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.849268] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8581f1eb-fc07-4c41-aad0-466056835d61 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.864435] env[62914]: DEBUG nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 563.869937] env[62914]: DEBUG nova.compute.provider_tree [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 564.214525] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.375040] env[62914]: INFO nova.virt.block_device [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Booting with volume 87946414-ae37-4cfd-b8b8-fd2a7c3942c4 at /dev/sda [ 564.378374] env[62914]: DEBUG nova.scheduler.client.report [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 564.492584] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7cb4e4a7-148b-4b35-ae96-ea91ef25cad5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.503936] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c1e26f-e800-4a50-80ab-5a4bd0240f9a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.527122] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4444a4a5-17b5-4b2c-a323-0b269461a6f6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.535087] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b25f4be8-a6c5-4c06-b897-c52cde66b74a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.564568] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f58dce0-9255-4b32-a5e6-3978567a792d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.571715] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606a57a6-24d5-4b7c-b5aa-9b318cd98620 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.586634] env[62914]: DEBUG nova.virt.block_device [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Updating existing volume attachment record: f6fdbab8-979b-4493-ae10-99ff2a91783c {{(pid=62914) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 564.782175] env[62914]: DEBUG nova.network.neutron [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Successfully created port: 8f6beb45-a893-4500-bfb4-d92c2f46c170 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 564.885317] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.045s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.885955] env[62914]: ERROR nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0451d3e7-8933-4ae7-ad82-9ffde652798f, please check neutron logs for more information. [ 564.885955] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Traceback (most recent call last): [ 564.885955] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 564.885955] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] self.driver.spawn(context, instance, image_meta, [ 564.885955] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 564.885955] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] self._vmops.spawn(context, instance, image_meta, injected_files, [ 564.885955] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 564.885955] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] vm_ref = self.build_virtual_machine(instance, [ 564.885955] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 564.885955] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] vif_infos = vmwarevif.get_vif_info(self._session, [ 564.885955] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 564.886312] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] for vif in network_info: [ 564.886312] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 564.886312] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] return self._sync_wrapper(fn, *args, **kwargs) [ 564.886312] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 564.886312] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] self.wait() [ 564.886312] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 564.886312] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] self[:] = self._gt.wait() [ 564.886312] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 564.886312] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] return self._exit_event.wait() [ 564.886312] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 564.886312] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] result = hub.switch() [ 564.886312] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 564.886312] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] return self.greenlet.switch() [ 564.887713] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.887713] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] result = function(*args, **kwargs) [ 564.887713] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 564.887713] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] return func(*args, **kwargs) [ 564.887713] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 564.887713] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] raise e [ 564.887713] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.887713] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] nwinfo = self.network_api.allocate_for_instance( [ 564.887713] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 564.887713] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] created_port_ids = self._update_ports_for_instance( [ 564.887713] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 564.887713] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] with excutils.save_and_reraise_exception(): [ 564.887713] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.888082] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] self.force_reraise() [ 564.888082] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.888082] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] raise self.value [ 564.888082] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 564.888082] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] updated_port = self._update_port( [ 564.888082] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.888082] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] _ensure_no_port_binding_failure(port) [ 564.888082] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.888082] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] raise exception.PortBindingFailed(port_id=port['id']) [ 564.888082] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] nova.exception.PortBindingFailed: Binding failed for port 0451d3e7-8933-4ae7-ad82-9ffde652798f, please check neutron logs for more information. [ 564.888082] env[62914]: ERROR nova.compute.manager [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] [ 564.888382] env[62914]: DEBUG nova.compute.utils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Binding failed for port 0451d3e7-8933-4ae7-ad82-9ffde652798f, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 564.888382] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.632s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.892153] env[62914]: INFO nova.compute.claims [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 564.895306] env[62914]: DEBUG nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Build of instance df6047b8-8e4b-48e2-ae54-f5fda0aa77da was re-scheduled: Binding failed for port 0451d3e7-8933-4ae7-ad82-9ffde652798f, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 564.895306] env[62914]: DEBUG nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 564.895306] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "refresh_cache-df6047b8-8e4b-48e2-ae54-f5fda0aa77da" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.895306] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "refresh_cache-df6047b8-8e4b-48e2-ae54-f5fda0aa77da" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.895496] env[62914]: DEBUG nova.network.neutron [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 564.977964] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Acquiring lock "8bda7658-2b86-4869-89e2-b10d7119ae18" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.978218] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Lock "8bda7658-2b86-4869-89e2-b10d7119ae18" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.437161] env[62914]: DEBUG nova.network.neutron [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.766157] env[62914]: DEBUG nova.network.neutron [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.272117] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "refresh_cache-df6047b8-8e4b-48e2-ae54-f5fda0aa77da" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.272117] env[62914]: DEBUG nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 566.272117] env[62914]: DEBUG nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 566.272117] env[62914]: DEBUG nova.network.neutron [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 566.319776] env[62914]: DEBUG nova.network.neutron [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.330528] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c149c932-2200-4047-b134-cffd31813812 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.339955] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbaceca0-72fd-4a72-ad90-42614e4fc348 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.378956] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2470addf-6df7-405b-9fc0-a789a5a63047 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.389248] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a9d2b9-cf31-402b-a010-9c2288a65fd6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.402916] env[62914]: DEBUG nova.compute.provider_tree [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 566.754593] env[62914]: DEBUG nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 566.754593] env[62914]: DEBUG nova.virt.hardware [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 566.754593] env[62914]: DEBUG nova.virt.hardware [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 566.754593] env[62914]: DEBUG nova.virt.hardware [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 566.754842] env[62914]: DEBUG nova.virt.hardware [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 566.754842] env[62914]: DEBUG nova.virt.hardware [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 566.754842] env[62914]: DEBUG nova.virt.hardware [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 566.754842] env[62914]: DEBUG nova.virt.hardware [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 566.754842] env[62914]: DEBUG nova.virt.hardware [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 566.754998] env[62914]: DEBUG nova.virt.hardware [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 566.754998] env[62914]: DEBUG nova.virt.hardware [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 566.754998] env[62914]: DEBUG nova.virt.hardware [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 566.755133] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2d1f90-d72e-4914-a7fd-1b4ccb1882af {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.763545] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e02627e-d583-49f4-b2ff-fdac2dc1a0de {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.827323] env[62914]: DEBUG nova.network.neutron [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.906225] env[62914]: DEBUG nova.scheduler.client.report [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.976044] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Acquiring lock "15938dfe-a697-450b-8c88-ff4ef703ee1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.976362] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Lock "15938dfe-a697-450b-8c88-ff4ef703ee1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.334053] env[62914]: INFO nova.compute.manager [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: df6047b8-8e4b-48e2-ae54-f5fda0aa77da] Took 1.06 seconds to deallocate network for instance. [ 567.416330] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.528s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.416760] env[62914]: DEBUG nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 567.419718] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.602s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.419915] env[62914]: DEBUG nova.objects.instance [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62914) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 567.921498] env[62914]: DEBUG nova.compute.utils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 567.923602] env[62914]: DEBUG nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 567.923779] env[62914]: DEBUG nova.network.neutron [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 568.033539] env[62914]: DEBUG nova.policy [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd510ff90e25945dcacf62800cd2aa809', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '951507a49dcc46f08f7e610280f6beb4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 568.378226] env[62914]: INFO nova.scheduler.client.report [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Deleted allocations for instance df6047b8-8e4b-48e2-ae54-f5fda0aa77da [ 568.426336] env[62914]: DEBUG nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 568.433149] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc549d1b-66c3-43e3-ab69-dda66ce73d13 tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.434284] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.525s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.582229] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Acquiring lock "91328e42-ba3c-4d63-a805-973c2bfa2afb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.582229] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Lock "91328e42-ba3c-4d63-a805-973c2bfa2afb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.586411] env[62914]: ERROR nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8f6beb45-a893-4500-bfb4-d92c2f46c170, please check neutron logs for more information. [ 568.586411] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 568.586411] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.586411] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 568.586411] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.586411] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 568.586411] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.586411] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 568.586411] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.586411] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 568.586411] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.586411] env[62914]: ERROR nova.compute.manager raise self.value [ 568.586411] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.586411] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 568.586411] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.586411] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 568.587265] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.587265] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 568.587265] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8f6beb45-a893-4500-bfb4-d92c2f46c170, please check neutron logs for more information. [ 568.587265] env[62914]: ERROR nova.compute.manager [ 568.587265] env[62914]: Traceback (most recent call last): [ 568.587265] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 568.587265] env[62914]: listener.cb(fileno) [ 568.587265] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.587265] env[62914]: result = function(*args, **kwargs) [ 568.587265] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 568.587265] env[62914]: return func(*args, **kwargs) [ 568.587265] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.587265] env[62914]: raise e [ 568.587265] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.587265] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 568.587265] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.587265] env[62914]: created_port_ids = self._update_ports_for_instance( [ 568.587265] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.587265] env[62914]: with excutils.save_and_reraise_exception(): [ 568.587265] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.587265] env[62914]: self.force_reraise() [ 568.587265] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.587265] env[62914]: raise self.value [ 568.587265] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.587265] env[62914]: updated_port = self._update_port( [ 568.587265] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.587265] env[62914]: _ensure_no_port_binding_failure(port) [ 568.587265] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.587265] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 568.590153] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 8f6beb45-a893-4500-bfb4-d92c2f46c170, please check neutron logs for more information. [ 568.590153] env[62914]: Removing descriptor: 19 [ 568.592414] env[62914]: ERROR nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8f6beb45-a893-4500-bfb4-d92c2f46c170, please check neutron logs for more information. [ 568.592414] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Traceback (most recent call last): [ 568.592414] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 568.592414] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] yield resources [ 568.592414] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.592414] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] self.driver.spawn(context, instance, image_meta, [ 568.592414] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 568.592414] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.592414] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.592414] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] vm_ref = self.build_virtual_machine(instance, [ 568.592414] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.592917] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.592917] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.592917] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] for vif in network_info: [ 568.592917] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.592917] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] return self._sync_wrapper(fn, *args, **kwargs) [ 568.592917] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.592917] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] self.wait() [ 568.592917] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.592917] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] self[:] = self._gt.wait() [ 568.592917] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.592917] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] return self._exit_event.wait() [ 568.592917] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.592917] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] result = hub.switch() [ 568.594635] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.594635] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] return self.greenlet.switch() [ 568.594635] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.594635] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] result = function(*args, **kwargs) [ 568.594635] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 568.594635] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] return func(*args, **kwargs) [ 568.594635] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.594635] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] raise e [ 568.594635] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.594635] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] nwinfo = self.network_api.allocate_for_instance( [ 568.594635] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.594635] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] created_port_ids = self._update_ports_for_instance( [ 568.594635] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.595303] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] with excutils.save_and_reraise_exception(): [ 568.595303] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.595303] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] self.force_reraise() [ 568.595303] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.595303] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] raise self.value [ 568.595303] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.595303] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] updated_port = self._update_port( [ 568.595303] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.595303] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] _ensure_no_port_binding_failure(port) [ 568.595303] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.595303] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] raise exception.PortBindingFailed(port_id=port['id']) [ 568.595303] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] nova.exception.PortBindingFailed: Binding failed for port 8f6beb45-a893-4500-bfb4-d92c2f46c170, please check neutron logs for more information. [ 568.595303] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] [ 568.596193] env[62914]: INFO nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Terminating instance [ 568.600703] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Acquiring lock "refresh_cache-4981a04f-98e6-474b-850a-634d977d9e42" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.600807] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Acquired lock "refresh_cache-4981a04f-98e6-474b-850a-634d977d9e42" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.600959] env[62914]: DEBUG nova.network.neutron [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 568.672450] env[62914]: DEBUG nova.compute.manager [req-7d6a7920-f480-42e8-afe3-c432a56d5cdc req-685e157b-628b-4343-b208-4e3b7c6cca83 service nova] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Received event network-changed-8f6beb45-a893-4500-bfb4-d92c2f46c170 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 568.672746] env[62914]: DEBUG nova.compute.manager [req-7d6a7920-f480-42e8-afe3-c432a56d5cdc req-685e157b-628b-4343-b208-4e3b7c6cca83 service nova] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Refreshing instance network info cache due to event network-changed-8f6beb45-a893-4500-bfb4-d92c2f46c170. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 568.673108] env[62914]: DEBUG oslo_concurrency.lockutils [req-7d6a7920-f480-42e8-afe3-c432a56d5cdc req-685e157b-628b-4343-b208-4e3b7c6cca83 service nova] Acquiring lock "refresh_cache-4981a04f-98e6-474b-850a-634d977d9e42" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.889222] env[62914]: DEBUG oslo_concurrency.lockutils [None req-394a1eb2-f74e-4895-b4cd-d844c7ac9202 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "df6047b8-8e4b-48e2-ae54-f5fda0aa77da" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.658s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.048544] env[62914]: DEBUG nova.network.neutron [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Successfully created port: 34b9a0ca-528b-46fc-8631-87bc3661e69b {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 569.244319] env[62914]: DEBUG nova.network.neutron [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.394544] env[62914]: DEBUG nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 569.416851] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a3d359-4b45-4d8b-8404-559490feff2d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.426482] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87a9cafc-db4c-46ac-be82-7499c69c8566 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.466271] env[62914]: DEBUG nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 569.470094] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-640d3d71-1507-477d-bd6b-3f7a35b655ee {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.483504] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e687ee-27ad-4045-8436-f3dd25d4c8a7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.502529] env[62914]: DEBUG nova.compute.provider_tree [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.515534] env[62914]: DEBUG nova.virt.hardware [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 569.515787] env[62914]: DEBUG nova.virt.hardware [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 569.516134] env[62914]: DEBUG nova.virt.hardware [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 569.516318] env[62914]: DEBUG nova.virt.hardware [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 569.516478] env[62914]: DEBUG nova.virt.hardware [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 569.516620] env[62914]: DEBUG nova.virt.hardware [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 569.516949] env[62914]: DEBUG nova.virt.hardware [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 569.517019] env[62914]: DEBUG nova.virt.hardware [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 569.517158] env[62914]: DEBUG nova.virt.hardware [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 569.517316] env[62914]: DEBUG nova.virt.hardware [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 569.518148] env[62914]: DEBUG nova.virt.hardware [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 569.518379] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a142a2a-0763-45fb-b740-2d4f64831ae1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.528367] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f68f29-3998-42ab-8571-2d0de1df67ba {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.599407] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Acquiring lock "0419073e-5340-4805-a1e0-417ccf5fdd30" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.600443] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Lock "0419073e-5340-4805-a1e0-417ccf5fdd30" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.644455] env[62914]: DEBUG nova.network.neutron [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.930016] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.010586] env[62914]: DEBUG nova.scheduler.client.report [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.149706] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Releasing lock "refresh_cache-4981a04f-98e6-474b-850a-634d977d9e42" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.150315] env[62914]: DEBUG nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 570.150660] env[62914]: DEBUG oslo_concurrency.lockutils [req-7d6a7920-f480-42e8-afe3-c432a56d5cdc req-685e157b-628b-4343-b208-4e3b7c6cca83 service nova] Acquired lock "refresh_cache-4981a04f-98e6-474b-850a-634d977d9e42" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.150836] env[62914]: DEBUG nova.network.neutron [req-7d6a7920-f480-42e8-afe3-c432a56d5cdc req-685e157b-628b-4343-b208-4e3b7c6cca83 service nova] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Refreshing network info cache for port 8f6beb45-a893-4500-bfb4-d92c2f46c170 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 570.153439] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f2604b0-2632-4012-93f2-0dd7e29aebff {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.166723] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1315ae-1b52-4a6b-9cd2-c795c55681e2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.195427] env[62914]: WARNING nova.virt.vmwareapi.driver [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 4981a04f-98e6-474b-850a-634d977d9e42 could not be found. [ 570.197231] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 570.197231] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7363eeab-bca6-4c2b-aab5-2064bc0827f1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.205402] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f81109-2ea9-41ee-8f12-1fc5833cba7e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.229346] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4981a04f-98e6-474b-850a-634d977d9e42 could not be found. [ 570.229451] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 570.229618] env[62914]: INFO nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Took 0.08 seconds to destroy the instance on the hypervisor. [ 570.229895] env[62914]: DEBUG oslo.service.loopingcall [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 570.230089] env[62914]: DEBUG nova.compute.manager [-] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 570.230401] env[62914]: DEBUG nova.network.neutron [-] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 570.270127] env[62914]: DEBUG nova.network.neutron [-] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.517759] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.083s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.518953] env[62914]: ERROR nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2dd6fb02-f6a2-42e3-987f-e9be0b492068, please check neutron logs for more information. [ 570.518953] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Traceback (most recent call last): [ 570.518953] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 570.518953] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] self.driver.spawn(context, instance, image_meta, [ 570.518953] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 570.518953] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] self._vmops.spawn(context, instance, image_meta, injected_files, [ 570.518953] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 570.518953] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] vm_ref = self.build_virtual_machine(instance, [ 570.518953] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 570.518953] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] vif_infos = vmwarevif.get_vif_info(self._session, [ 570.518953] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 570.519375] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] for vif in network_info: [ 570.519375] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 570.519375] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] return self._sync_wrapper(fn, *args, **kwargs) [ 570.519375] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 570.519375] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] self.wait() [ 570.519375] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 570.519375] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] self[:] = self._gt.wait() [ 570.519375] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 570.519375] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] return self._exit_event.wait() [ 570.519375] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 570.519375] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] result = hub.switch() [ 570.519375] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 570.519375] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] return self.greenlet.switch() [ 570.519819] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 570.519819] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] result = function(*args, **kwargs) [ 570.519819] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 570.519819] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] return func(*args, **kwargs) [ 570.519819] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 570.519819] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] raise e [ 570.519819] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 570.519819] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] nwinfo = self.network_api.allocate_for_instance( [ 570.519819] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 570.519819] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] created_port_ids = self._update_ports_for_instance( [ 570.519819] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 570.519819] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] with excutils.save_and_reraise_exception(): [ 570.519819] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 570.520226] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] self.force_reraise() [ 570.520226] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 570.520226] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] raise self.value [ 570.520226] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 570.520226] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] updated_port = self._update_port( [ 570.520226] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 570.520226] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] _ensure_no_port_binding_failure(port) [ 570.520226] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 570.520226] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] raise exception.PortBindingFailed(port_id=port['id']) [ 570.520226] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] nova.exception.PortBindingFailed: Binding failed for port 2dd6fb02-f6a2-42e3-987f-e9be0b492068, please check neutron logs for more information. [ 570.520226] env[62914]: ERROR nova.compute.manager [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] [ 570.520531] env[62914]: DEBUG nova.compute.utils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Binding failed for port 2dd6fb02-f6a2-42e3-987f-e9be0b492068, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 570.522404] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.525s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.525760] env[62914]: INFO nova.compute.claims [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 570.532860] env[62914]: DEBUG nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Build of instance 16b4e8bd-d84f-45a6-b531-3200ea545931 was re-scheduled: Binding failed for port 2dd6fb02-f6a2-42e3-987f-e9be0b492068, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 570.532860] env[62914]: DEBUG nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 570.532860] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Acquiring lock "refresh_cache-16b4e8bd-d84f-45a6-b531-3200ea545931" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.532860] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Acquired lock "refresh_cache-16b4e8bd-d84f-45a6-b531-3200ea545931" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.533093] env[62914]: DEBUG nova.network.neutron [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 570.675997] env[62914]: DEBUG nova.network.neutron [req-7d6a7920-f480-42e8-afe3-c432a56d5cdc req-685e157b-628b-4343-b208-4e3b7c6cca83 service nova] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.773151] env[62914]: DEBUG nova.network.neutron [-] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.806634] env[62914]: DEBUG nova.network.neutron [req-7d6a7920-f480-42e8-afe3-c432a56d5cdc req-685e157b-628b-4343-b208-4e3b7c6cca83 service nova] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.063074] env[62914]: DEBUG nova.network.neutron [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.185717] env[62914]: DEBUG nova.network.neutron [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.279444] env[62914]: INFO nova.compute.manager [-] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Took 1.05 seconds to deallocate network for instance. [ 571.309699] env[62914]: DEBUG oslo_concurrency.lockutils [req-7d6a7920-f480-42e8-afe3-c432a56d5cdc req-685e157b-628b-4343-b208-4e3b7c6cca83 service nova] Releasing lock "refresh_cache-4981a04f-98e6-474b-850a-634d977d9e42" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.372191] env[62914]: DEBUG nova.compute.manager [req-35e3472d-4090-4e6e-96bc-fa22e770d646 req-fc7b4ad4-6448-44dd-b9e7-ce586dd483ac service nova] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Received event network-vif-deleted-8f6beb45-a893-4500-bfb4-d92c2f46c170 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 571.415792] env[62914]: ERROR nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 34b9a0ca-528b-46fc-8631-87bc3661e69b, please check neutron logs for more information. [ 571.415792] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 571.415792] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.415792] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 571.415792] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.415792] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 571.415792] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.415792] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 571.415792] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.415792] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 571.415792] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.415792] env[62914]: ERROR nova.compute.manager raise self.value [ 571.415792] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.415792] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 571.415792] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.415792] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 571.416356] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.416356] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 571.416356] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 34b9a0ca-528b-46fc-8631-87bc3661e69b, please check neutron logs for more information. [ 571.416356] env[62914]: ERROR nova.compute.manager [ 571.416356] env[62914]: Traceback (most recent call last): [ 571.416356] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 571.416356] env[62914]: listener.cb(fileno) [ 571.416356] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.416356] env[62914]: result = function(*args, **kwargs) [ 571.416356] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 571.416356] env[62914]: return func(*args, **kwargs) [ 571.416356] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 571.416356] env[62914]: raise e [ 571.416356] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.416356] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 571.416356] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.416356] env[62914]: created_port_ids = self._update_ports_for_instance( [ 571.416356] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.416356] env[62914]: with excutils.save_and_reraise_exception(): [ 571.416356] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.416356] env[62914]: self.force_reraise() [ 571.416356] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.416356] env[62914]: raise self.value [ 571.416356] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.416356] env[62914]: updated_port = self._update_port( [ 571.416356] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.416356] env[62914]: _ensure_no_port_binding_failure(port) [ 571.416356] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.416356] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 571.417230] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 34b9a0ca-528b-46fc-8631-87bc3661e69b, please check neutron logs for more information. [ 571.417230] env[62914]: Removing descriptor: 15 [ 571.417230] env[62914]: ERROR nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 34b9a0ca-528b-46fc-8631-87bc3661e69b, please check neutron logs for more information. [ 571.417230] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Traceback (most recent call last): [ 571.417230] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 571.417230] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] yield resources [ 571.417230] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 571.417230] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] self.driver.spawn(context, instance, image_meta, [ 571.417230] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 571.417230] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 571.417230] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 571.417230] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] vm_ref = self.build_virtual_machine(instance, [ 571.417654] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 571.417654] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 571.417654] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 571.417654] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] for vif in network_info: [ 571.417654] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 571.417654] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] return self._sync_wrapper(fn, *args, **kwargs) [ 571.417654] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 571.417654] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] self.wait() [ 571.417654] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 571.417654] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] self[:] = self._gt.wait() [ 571.417654] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 571.417654] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] return self._exit_event.wait() [ 571.417654] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 571.418082] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] result = hub.switch() [ 571.418082] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 571.418082] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] return self.greenlet.switch() [ 571.418082] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.418082] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] result = function(*args, **kwargs) [ 571.418082] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 571.418082] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] return func(*args, **kwargs) [ 571.418082] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 571.418082] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] raise e [ 571.418082] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.418082] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] nwinfo = self.network_api.allocate_for_instance( [ 571.418082] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.418082] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] created_port_ids = self._update_ports_for_instance( [ 571.418485] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.418485] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] with excutils.save_and_reraise_exception(): [ 571.418485] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.418485] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] self.force_reraise() [ 571.418485] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.418485] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] raise self.value [ 571.418485] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.418485] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] updated_port = self._update_port( [ 571.418485] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.418485] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] _ensure_no_port_binding_failure(port) [ 571.418485] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.418485] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] raise exception.PortBindingFailed(port_id=port['id']) [ 571.418839] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] nova.exception.PortBindingFailed: Binding failed for port 34b9a0ca-528b-46fc-8631-87bc3661e69b, please check neutron logs for more information. [ 571.418839] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] [ 571.418839] env[62914]: INFO nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Terminating instance [ 571.421504] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Acquiring lock "refresh_cache-f1cee1cb-aceb-47e8-b2da-b94563f922e2" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.421504] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Acquired lock "refresh_cache-f1cee1cb-aceb-47e8-b2da-b94563f922e2" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.421639] env[62914]: DEBUG nova.network.neutron [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 571.466153] env[62914]: DEBUG nova.compute.manager [req-3f86e1d2-2516-41d9-8012-141c856bfff9 req-b9df2906-f50f-4964-a669-b9ccb2ec2892 service nova] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Received event network-changed-34b9a0ca-528b-46fc-8631-87bc3661e69b {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 571.466247] env[62914]: DEBUG nova.compute.manager [req-3f86e1d2-2516-41d9-8012-141c856bfff9 req-b9df2906-f50f-4964-a669-b9ccb2ec2892 service nova] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Refreshing instance network info cache due to event network-changed-34b9a0ca-528b-46fc-8631-87bc3661e69b. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 571.466372] env[62914]: DEBUG oslo_concurrency.lockutils [req-3f86e1d2-2516-41d9-8012-141c856bfff9 req-b9df2906-f50f-4964-a669-b9ccb2ec2892 service nova] Acquiring lock "refresh_cache-f1cee1cb-aceb-47e8-b2da-b94563f922e2" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.692027] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Releasing lock "refresh_cache-16b4e8bd-d84f-45a6-b531-3200ea545931" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.692027] env[62914]: DEBUG nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 571.692027] env[62914]: DEBUG nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 571.692027] env[62914]: DEBUG nova.network.neutron [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 571.723733] env[62914]: DEBUG nova.network.neutron [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.870149] env[62914]: INFO nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Took 0.59 seconds to detach 1 volumes for instance. [ 571.875134] env[62914]: DEBUG nova.compute.claims [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 571.875134] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.947620] env[62914]: DEBUG nova.network.neutron [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.967808] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d129cc0b-c36c-449f-b1d9-226ded884808 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.977675] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a854ea7-fb65-4e03-b7ba-7a2dc51eb704 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.017262] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba55f1d7-5dbe-4716-801d-415afaff10f7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.025746] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-570ae5da-49f4-4218-beaf-92640571ae2c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.041260] env[62914]: DEBUG nova.compute.provider_tree [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.044516] env[62914]: DEBUG nova.network.neutron [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.227095] env[62914]: DEBUG nova.network.neutron [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.470566] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Acquiring lock "c10d840e-ce05-4af3-ba2d-6214eefd8783" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.470989] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Lock "c10d840e-ce05-4af3-ba2d-6214eefd8783" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.547450] env[62914]: DEBUG nova.scheduler.client.report [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.549516] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Releasing lock "refresh_cache-f1cee1cb-aceb-47e8-b2da-b94563f922e2" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.549711] env[62914]: DEBUG nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 572.549795] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 572.553080] env[62914]: DEBUG oslo_concurrency.lockutils [req-3f86e1d2-2516-41d9-8012-141c856bfff9 req-b9df2906-f50f-4964-a669-b9ccb2ec2892 service nova] Acquired lock "refresh_cache-f1cee1cb-aceb-47e8-b2da-b94563f922e2" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.553080] env[62914]: DEBUG nova.network.neutron [req-3f86e1d2-2516-41d9-8012-141c856bfff9 req-b9df2906-f50f-4964-a669-b9ccb2ec2892 service nova] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Refreshing network info cache for port 34b9a0ca-528b-46fc-8631-87bc3661e69b {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 572.553080] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7bace25f-33f9-4e19-b1fa-90d701a66f3c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.564301] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd9f227-4106-4752-b0f4-747c9160553b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.587734] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f1cee1cb-aceb-47e8-b2da-b94563f922e2 could not be found. [ 572.588014] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 572.588225] env[62914]: INFO nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 572.588967] env[62914]: DEBUG oslo.service.loopingcall [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 572.588967] env[62914]: DEBUG nova.compute.manager [-] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 572.588967] env[62914]: DEBUG nova.network.neutron [-] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 572.615847] env[62914]: DEBUG nova.network.neutron [-] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.730716] env[62914]: INFO nova.compute.manager [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] [instance: 16b4e8bd-d84f-45a6-b531-3200ea545931] Took 1.04 seconds to deallocate network for instance. [ 572.957534] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "c5aaadb0-f9a1-4af9-b882-86d847185a19" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.957806] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "c5aaadb0-f9a1-4af9-b882-86d847185a19" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.054387] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.054955] env[62914]: DEBUG nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 573.061070] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.633s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.086449] env[62914]: DEBUG nova.network.neutron [req-3f86e1d2-2516-41d9-8012-141c856bfff9 req-b9df2906-f50f-4964-a669-b9ccb2ec2892 service nova] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.118655] env[62914]: DEBUG nova.network.neutron [-] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.188792] env[62914]: DEBUG nova.network.neutron [req-3f86e1d2-2516-41d9-8012-141c856bfff9 req-b9df2906-f50f-4964-a669-b9ccb2ec2892 service nova] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.430317] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Acquiring lock "62cfca0a-e8c6-4f94-8e6f-1136cfcf3089" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.430654] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Lock "62cfca0a-e8c6-4f94-8e6f-1136cfcf3089" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.560042] env[62914]: DEBUG nova.compute.utils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 573.560907] env[62914]: DEBUG nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 573.561080] env[62914]: DEBUG nova.network.neutron [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 573.623168] env[62914]: INFO nova.compute.manager [-] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Took 1.03 seconds to deallocate network for instance. [ 573.627244] env[62914]: DEBUG nova.compute.claims [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 573.628569] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.651072] env[62914]: DEBUG nova.policy [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '86ccf67d949f4586b023ee45005e22dc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d0af14f259b4f748bdf079156c5561f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 573.692386] env[62914]: DEBUG oslo_concurrency.lockutils [req-3f86e1d2-2516-41d9-8012-141c856bfff9 req-b9df2906-f50f-4964-a669-b9ccb2ec2892 service nova] Releasing lock "refresh_cache-f1cee1cb-aceb-47e8-b2da-b94563f922e2" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.762894] env[62914]: INFO nova.scheduler.client.report [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Deleted allocations for instance 16b4e8bd-d84f-45a6-b531-3200ea545931 [ 573.985240] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8083eb00-3ca8-45e1-803a-c95211d76971 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.993459] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0cda07-bad3-4c6b-a420-02f830fdfe4c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.027474] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf9ec55d-ec2e-4cf3-9888-604c18601726 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.030659] env[62914]: DEBUG nova.compute.manager [req-150452fd-11a3-487c-941b-f2ebff794073 req-b6afa773-5083-452f-8e60-715d16543c8f service nova] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Received event network-vif-deleted-34b9a0ca-528b-46fc-8631-87bc3661e69b {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 574.037403] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7747c56b-61ff-4b4e-9110-67b0ce3d6ad6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.053226] env[62914]: DEBUG nova.compute.provider_tree [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.064770] env[62914]: DEBUG nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 574.216889] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Acquiring lock "294e3321-e826-4a56-bbd0-893261d65283" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.217245] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Lock "294e3321-e826-4a56-bbd0-893261d65283" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.252439] env[62914]: DEBUG nova.network.neutron [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Successfully created port: 50ee8c92-f158-4391-abba-a9828767fd9b {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 574.277765] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f74ac90d-a3af-4148-b82e-bc4505a44ab9 tempest-ServerExternalEventsTest-1853569396 tempest-ServerExternalEventsTest-1853569396-project-member] Lock "16b4e8bd-d84f-45a6-b531-3200ea545931" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.650s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.559623] env[62914]: DEBUG nova.scheduler.client.report [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.780882] env[62914]: DEBUG nova.compute.manager [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 575.065602] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.006s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.066303] env[62914]: ERROR nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fe1003d5-35d9-4496-87f7-dcbf94483f17, please check neutron logs for more information. [ 575.066303] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] Traceback (most recent call last): [ 575.066303] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.066303] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] self.driver.spawn(context, instance, image_meta, [ 575.066303] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 575.066303] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.066303] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.066303] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] vm_ref = self.build_virtual_machine(instance, [ 575.066303] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.066303] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.066303] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.066740] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] for vif in network_info: [ 575.066740] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.066740] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] return self._sync_wrapper(fn, *args, **kwargs) [ 575.066740] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.066740] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] self.wait() [ 575.066740] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.066740] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] self[:] = self._gt.wait() [ 575.066740] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.066740] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] return self._exit_event.wait() [ 575.066740] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.066740] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] result = hub.switch() [ 575.066740] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.066740] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] return self.greenlet.switch() [ 575.067151] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.067151] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] result = function(*args, **kwargs) [ 575.067151] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 575.067151] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] return func(*args, **kwargs) [ 575.067151] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.067151] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] raise e [ 575.067151] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.067151] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] nwinfo = self.network_api.allocate_for_instance( [ 575.067151] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.067151] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] created_port_ids = self._update_ports_for_instance( [ 575.067151] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.067151] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] with excutils.save_and_reraise_exception(): [ 575.067151] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.067608] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] self.force_reraise() [ 575.067608] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.067608] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] raise self.value [ 575.067608] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.067608] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] updated_port = self._update_port( [ 575.067608] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.067608] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] _ensure_no_port_binding_failure(port) [ 575.067608] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.067608] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] raise exception.PortBindingFailed(port_id=port['id']) [ 575.067608] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] nova.exception.PortBindingFailed: Binding failed for port fe1003d5-35d9-4496-87f7-dcbf94483f17, please check neutron logs for more information. [ 575.067608] env[62914]: ERROR nova.compute.manager [instance: a8761350-3044-4822-8221-ed108e21cd62] [ 575.067935] env[62914]: DEBUG nova.compute.utils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Binding failed for port fe1003d5-35d9-4496-87f7-dcbf94483f17, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 575.070722] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.628s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.074037] env[62914]: DEBUG nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Build of instance a8761350-3044-4822-8221-ed108e21cd62 was re-scheduled: Binding failed for port fe1003d5-35d9-4496-87f7-dcbf94483f17, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 575.076036] env[62914]: DEBUG nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 575.076912] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Acquiring lock "refresh_cache-a8761350-3044-4822-8221-ed108e21cd62" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.076912] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Acquired lock "refresh_cache-a8761350-3044-4822-8221-ed108e21cd62" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.077880] env[62914]: DEBUG nova.network.neutron [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 575.079718] env[62914]: DEBUG nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 575.129155] env[62914]: DEBUG nova.virt.hardware [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 575.129770] env[62914]: DEBUG nova.virt.hardware [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 575.129770] env[62914]: DEBUG nova.virt.hardware [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 575.129770] env[62914]: DEBUG nova.virt.hardware [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 575.129940] env[62914]: DEBUG nova.virt.hardware [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 575.129940] env[62914]: DEBUG nova.virt.hardware [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 575.130211] env[62914]: DEBUG nova.virt.hardware [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 575.130304] env[62914]: DEBUG nova.virt.hardware [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 575.130475] env[62914]: DEBUG nova.virt.hardware [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 575.131770] env[62914]: DEBUG nova.virt.hardware [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 575.131770] env[62914]: DEBUG nova.virt.hardware [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 575.131962] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ac8811-39bd-4525-9aba-e68b89a6173b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.140891] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d6273b1-452f-4429-b22d-d3ff77195518 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.163051] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Acquiring lock "0ac803ee-6303-463f-91e8-8263c0219b0a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.163051] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Lock "0ac803ee-6303-463f-91e8-8263c0219b0a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.303395] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.607454] env[62914]: DEBUG nova.network.neutron [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.662199] env[62914]: DEBUG nova.network.neutron [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.082928] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0760457c-684f-40ef-8b92-119fab2a05be {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.091611] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54fad1b-8e57-4c41-becb-d6e01beb1c6c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.128191] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8198a5e7-33b8-4466-a0f4-a22ee9fb8b38 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.137469] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acee6da1-a24b-4cf1-a184-6ad14431b7fe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.153226] env[62914]: DEBUG nova.compute.provider_tree [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.165213] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Releasing lock "refresh_cache-a8761350-3044-4822-8221-ed108e21cd62" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.165452] env[62914]: DEBUG nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 576.165611] env[62914]: DEBUG nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 576.165770] env[62914]: DEBUG nova.network.neutron [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 576.184993] env[62914]: DEBUG nova.network.neutron [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.294030] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Acquiring lock "3ce62ec7-6274-441f-989b-6f27a64b0c18" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.294273] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Lock "3ce62ec7-6274-441f-989b-6f27a64b0c18" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.658028] env[62914]: DEBUG nova.scheduler.client.report [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 576.684371] env[62914]: DEBUG nova.compute.manager [req-e1456ba8-3988-4e6a-857e-ff14eede7f0b req-3b2a8201-2cd6-4596-afb4-61a332c5f8a5 service nova] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Received event network-changed-50ee8c92-f158-4391-abba-a9828767fd9b {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 576.684569] env[62914]: DEBUG nova.compute.manager [req-e1456ba8-3988-4e6a-857e-ff14eede7f0b req-3b2a8201-2cd6-4596-afb4-61a332c5f8a5 service nova] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Refreshing instance network info cache due to event network-changed-50ee8c92-f158-4391-abba-a9828767fd9b. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 576.684749] env[62914]: DEBUG oslo_concurrency.lockutils [req-e1456ba8-3988-4e6a-857e-ff14eede7f0b req-3b2a8201-2cd6-4596-afb4-61a332c5f8a5 service nova] Acquiring lock "refresh_cache-6ec0c746-2bb8-4387-8b31-b49b0d708614" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.684884] env[62914]: DEBUG oslo_concurrency.lockutils [req-e1456ba8-3988-4e6a-857e-ff14eede7f0b req-3b2a8201-2cd6-4596-afb4-61a332c5f8a5 service nova] Acquired lock "refresh_cache-6ec0c746-2bb8-4387-8b31-b49b0d708614" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.685042] env[62914]: DEBUG nova.network.neutron [req-e1456ba8-3988-4e6a-857e-ff14eede7f0b req-3b2a8201-2cd6-4596-afb4-61a332c5f8a5 service nova] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Refreshing network info cache for port 50ee8c92-f158-4391-abba-a9828767fd9b {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 576.687131] env[62914]: ERROR nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 50ee8c92-f158-4391-abba-a9828767fd9b, please check neutron logs for more information. [ 576.687131] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 576.687131] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.687131] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 576.687131] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.687131] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 576.687131] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.687131] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 576.687131] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.687131] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 576.687131] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.687131] env[62914]: ERROR nova.compute.manager raise self.value [ 576.687131] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.687131] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 576.687131] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.687131] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 576.687714] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.687714] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 576.687714] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 50ee8c92-f158-4391-abba-a9828767fd9b, please check neutron logs for more information. [ 576.687714] env[62914]: ERROR nova.compute.manager [ 576.687714] env[62914]: Traceback (most recent call last): [ 576.687714] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 576.687714] env[62914]: listener.cb(fileno) [ 576.687714] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.687714] env[62914]: result = function(*args, **kwargs) [ 576.687714] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 576.687714] env[62914]: return func(*args, **kwargs) [ 576.687714] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 576.687714] env[62914]: raise e [ 576.687714] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.687714] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 576.687714] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.687714] env[62914]: created_port_ids = self._update_ports_for_instance( [ 576.687714] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.687714] env[62914]: with excutils.save_and_reraise_exception(): [ 576.687714] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.687714] env[62914]: self.force_reraise() [ 576.687714] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.687714] env[62914]: raise self.value [ 576.687714] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.687714] env[62914]: updated_port = self._update_port( [ 576.687714] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.687714] env[62914]: _ensure_no_port_binding_failure(port) [ 576.687714] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.687714] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 576.688726] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 50ee8c92-f158-4391-abba-a9828767fd9b, please check neutron logs for more information. [ 576.688726] env[62914]: Removing descriptor: 15 [ 576.688726] env[62914]: DEBUG nova.network.neutron [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.689517] env[62914]: ERROR nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 50ee8c92-f158-4391-abba-a9828767fd9b, please check neutron logs for more information. [ 576.689517] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Traceback (most recent call last): [ 576.689517] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 576.689517] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] yield resources [ 576.689517] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 576.689517] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] self.driver.spawn(context, instance, image_meta, [ 576.689517] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 576.689517] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.689517] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.689517] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] vm_ref = self.build_virtual_machine(instance, [ 576.689517] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.690171] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.690171] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.690171] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] for vif in network_info: [ 576.690171] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 576.690171] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] return self._sync_wrapper(fn, *args, **kwargs) [ 576.690171] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 576.690171] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] self.wait() [ 576.690171] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 576.690171] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] self[:] = self._gt.wait() [ 576.690171] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.690171] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] return self._exit_event.wait() [ 576.690171] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.690171] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] result = hub.switch() [ 576.690749] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.690749] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] return self.greenlet.switch() [ 576.690749] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.690749] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] result = function(*args, **kwargs) [ 576.690749] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 576.690749] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] return func(*args, **kwargs) [ 576.690749] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 576.690749] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] raise e [ 576.690749] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.690749] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] nwinfo = self.network_api.allocate_for_instance( [ 576.690749] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.690749] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] created_port_ids = self._update_ports_for_instance( [ 576.690749] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.691698] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] with excutils.save_and_reraise_exception(): [ 576.691698] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.691698] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] self.force_reraise() [ 576.691698] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.691698] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] raise self.value [ 576.691698] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.691698] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] updated_port = self._update_port( [ 576.691698] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.691698] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] _ensure_no_port_binding_failure(port) [ 576.691698] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.691698] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] raise exception.PortBindingFailed(port_id=port['id']) [ 576.691698] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] nova.exception.PortBindingFailed: Binding failed for port 50ee8c92-f158-4391-abba-a9828767fd9b, please check neutron logs for more information. [ 576.691698] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] [ 576.692314] env[62914]: INFO nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Terminating instance [ 576.692314] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Acquiring lock "refresh_cache-6ec0c746-2bb8-4387-8b31-b49b0d708614" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.161845] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.091s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.162437] env[62914]: ERROR nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8f3888a9-3b63-4288-99ab-32523f8483a2, please check neutron logs for more information. [ 577.162437] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Traceback (most recent call last): [ 577.162437] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.162437] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] self.driver.spawn(context, instance, image_meta, [ 577.162437] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 577.162437] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.162437] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.162437] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] vm_ref = self.build_virtual_machine(instance, [ 577.162437] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.162437] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.162437] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.162806] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] for vif in network_info: [ 577.162806] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.162806] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] return self._sync_wrapper(fn, *args, **kwargs) [ 577.162806] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.162806] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] self.wait() [ 577.162806] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.162806] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] self[:] = self._gt.wait() [ 577.162806] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.162806] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] return self._exit_event.wait() [ 577.162806] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.162806] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] result = hub.switch() [ 577.162806] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.162806] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] return self.greenlet.switch() [ 577.163318] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.163318] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] result = function(*args, **kwargs) [ 577.163318] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 577.163318] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] return func(*args, **kwargs) [ 577.163318] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.163318] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] raise e [ 577.163318] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.163318] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] nwinfo = self.network_api.allocate_for_instance( [ 577.163318] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 577.163318] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] created_port_ids = self._update_ports_for_instance( [ 577.163318] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 577.163318] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] with excutils.save_and_reraise_exception(): [ 577.163318] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.163765] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] self.force_reraise() [ 577.163765] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.163765] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] raise self.value [ 577.163765] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 577.163765] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] updated_port = self._update_port( [ 577.163765] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.163765] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] _ensure_no_port_binding_failure(port) [ 577.163765] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.163765] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] raise exception.PortBindingFailed(port_id=port['id']) [ 577.163765] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] nova.exception.PortBindingFailed: Binding failed for port 8f3888a9-3b63-4288-99ab-32523f8483a2, please check neutron logs for more information. [ 577.163765] env[62914]: ERROR nova.compute.manager [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] [ 577.164191] env[62914]: DEBUG nova.compute.utils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Binding failed for port 8f3888a9-3b63-4288-99ab-32523f8483a2, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 577.168585] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.651s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.168585] env[62914]: INFO nova.compute.claims [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 577.172038] env[62914]: DEBUG nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Build of instance 856b3816-d4f8-4c9e-b74c-ea15d67976ef was re-scheduled: Binding failed for port 8f3888a9-3b63-4288-99ab-32523f8483a2, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 577.172038] env[62914]: DEBUG nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 577.172038] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Acquiring lock "refresh_cache-856b3816-d4f8-4c9e-b74c-ea15d67976ef" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.172038] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Acquired lock "refresh_cache-856b3816-d4f8-4c9e-b74c-ea15d67976ef" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.172290] env[62914]: DEBUG nova.network.neutron [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 577.191612] env[62914]: INFO nova.compute.manager [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: a8761350-3044-4822-8221-ed108e21cd62] Took 1.03 seconds to deallocate network for instance. [ 577.210892] env[62914]: DEBUG nova.network.neutron [req-e1456ba8-3988-4e6a-857e-ff14eede7f0b req-3b2a8201-2cd6-4596-afb4-61a332c5f8a5 service nova] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.422539] env[62914]: DEBUG nova.network.neutron [req-e1456ba8-3988-4e6a-857e-ff14eede7f0b req-3b2a8201-2cd6-4596-afb4-61a332c5f8a5 service nova] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.712378] env[62914]: DEBUG nova.network.neutron [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.927222] env[62914]: DEBUG oslo_concurrency.lockutils [req-e1456ba8-3988-4e6a-857e-ff14eede7f0b req-3b2a8201-2cd6-4596-afb4-61a332c5f8a5 service nova] Releasing lock "refresh_cache-6ec0c746-2bb8-4387-8b31-b49b0d708614" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.927222] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Acquired lock "refresh_cache-6ec0c746-2bb8-4387-8b31-b49b0d708614" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.927222] env[62914]: DEBUG nova.network.neutron [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 577.939284] env[62914]: DEBUG nova.network.neutron [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.246513] env[62914]: INFO nova.scheduler.client.report [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Deleted allocations for instance a8761350-3044-4822-8221-ed108e21cd62 [ 578.420323] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Acquiring lock "e24c8bf2-1616-4a5a-b6ce-0c256e78b213" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.423028] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Lock "e24c8bf2-1616-4a5a-b6ce-0c256e78b213" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.444070] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Releasing lock "refresh_cache-856b3816-d4f8-4c9e-b74c-ea15d67976ef" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.444070] env[62914]: DEBUG nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 578.444070] env[62914]: DEBUG nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 578.444070] env[62914]: DEBUG nova.network.neutron [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 578.470200] env[62914]: DEBUG nova.network.neutron [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.471698] env[62914]: DEBUG nova.network.neutron [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.636371] env[62914]: DEBUG nova.network.neutron [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.752568] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d05ad2e-c002-4d83-b4ca-ba48401650c7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.763436] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51612b59-80af-4414-8606-495cb6ccf5af tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Lock "a8761350-3044-4822-8221-ed108e21cd62" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.595s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.764715] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6ce20e-bf4d-49b6-9fdb-ca631e70773a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.802824] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f06cb93-7de2-481d-abce-73348ed3bc89 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.811454] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432b039a-16a8-4020-8e6b-def917d7d997 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.826183] env[62914]: DEBUG nova.compute.provider_tree [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.974945] env[62914]: DEBUG nova.network.neutron [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.058229] env[62914]: DEBUG nova.compute.manager [req-eac56af8-311e-4333-bf5f-2764b3f588dc req-9f0306f6-d730-494c-bb48-13ccad9fbede service nova] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Received event network-vif-deleted-50ee8c92-f158-4391-abba-a9828767fd9b {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.139223] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Releasing lock "refresh_cache-6ec0c746-2bb8-4387-8b31-b49b0d708614" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.139709] env[62914]: DEBUG nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 579.140186] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 579.140550] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8c12fb3-87ed-4284-a935-52bf05437808 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.152577] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd88c496-a925-4843-9dda-087b71c60c8f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.177851] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6ec0c746-2bb8-4387-8b31-b49b0d708614 could not be found. [ 579.178108] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 579.178421] env[62914]: INFO nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Took 0.04 seconds to destroy the instance on the hypervisor. [ 579.178795] env[62914]: DEBUG oslo.service.loopingcall [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 579.179085] env[62914]: DEBUG nova.compute.manager [-] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 579.179195] env[62914]: DEBUG nova.network.neutron [-] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 579.203848] env[62914]: DEBUG nova.network.neutron [-] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.269621] env[62914]: DEBUG nova.compute.manager [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 579.329069] env[62914]: DEBUG nova.scheduler.client.report [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.478404] env[62914]: INFO nova.compute.manager [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] [instance: 856b3816-d4f8-4c9e-b74c-ea15d67976ef] Took 1.04 seconds to deallocate network for instance. [ 579.707105] env[62914]: DEBUG nova.network.neutron [-] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.806588] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.834960] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.670s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.837166] env[62914]: DEBUG nova.compute.manager [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 579.842100] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.208s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.843199] env[62914]: DEBUG nova.objects.instance [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Lazy-loading 'resources' on Instance uuid d5c90d13-41bc-4cca-abf3-5a5b13171a6f {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 580.210272] env[62914]: INFO nova.compute.manager [-] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Took 1.03 seconds to deallocate network for instance. [ 580.213571] env[62914]: DEBUG nova.compute.claims [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 580.213745] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.346302] env[62914]: DEBUG nova.compute.utils [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 580.351397] env[62914]: DEBUG nova.compute.manager [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Not allocating networking since 'none' was specified. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 580.520080] env[62914]: INFO nova.scheduler.client.report [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Deleted allocations for instance 856b3816-d4f8-4c9e-b74c-ea15d67976ef [ 580.836683] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6634f4fa-6f2e-4ed6-8bfa-3a34d9225c8e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.845752] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c785a79f-c4ee-4f3b-a991-d46fa3687206 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.851754] env[62914]: DEBUG nova.compute.manager [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 580.879438] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-403232e4-4c6b-4eee-899e-c4f9ac84f3c9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.887269] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2355927c-fa17-47f4-83b6-8b5f69d577f6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.900884] env[62914]: DEBUG nova.compute.provider_tree [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.035108] env[62914]: DEBUG oslo_concurrency.lockutils [None req-016295d3-224d-42e1-8d2e-fedad814dc4c tempest-VolumesAssistedSnapshotsTest-1322186147 tempest-VolumesAssistedSnapshotsTest-1322186147-project-member] Lock "856b3816-d4f8-4c9e-b74c-ea15d67976ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.974s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.405881] env[62914]: DEBUG nova.scheduler.client.report [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 581.542847] env[62914]: DEBUG nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 581.835961] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Acquiring lock "ad283bc7-b422-4836-a503-d3d9ae4fe3c1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.836227] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Lock "ad283bc7-b422-4836-a503-d3d9ae4fe3c1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.864536] env[62914]: DEBUG nova.compute.manager [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 581.906268] env[62914]: DEBUG nova.virt.hardware [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 581.906268] env[62914]: DEBUG nova.virt.hardware [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 581.906459] env[62914]: DEBUG nova.virt.hardware [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 581.906591] env[62914]: DEBUG nova.virt.hardware [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 581.906714] env[62914]: DEBUG nova.virt.hardware [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 581.906800] env[62914]: DEBUG nova.virt.hardware [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 581.907011] env[62914]: DEBUG nova.virt.hardware [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 581.907164] env[62914]: DEBUG nova.virt.hardware [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 581.907320] env[62914]: DEBUG nova.virt.hardware [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 581.907505] env[62914]: DEBUG nova.virt.hardware [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 581.907643] env[62914]: DEBUG nova.virt.hardware [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 581.908893] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b96f09be-aa00-465c-aa38-32223e246f65 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.915605] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.077s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.917708] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.353s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.917870] env[62914]: DEBUG nova.objects.instance [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62914) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 581.931706] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6102bd06-c4f0-49d1-a291-c22d57842581 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.948071] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Instance VIF info [] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 581.959062] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Creating folder: Project (2d7728b39d5f496196a09689732f21ab). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 581.960497] env[62914]: INFO nova.scheduler.client.report [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Deleted allocations for instance d5c90d13-41bc-4cca-abf3-5a5b13171a6f [ 581.961782] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b9421423-6bef-4839-97ae-14e011797bed {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.978272] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Created folder: Project (2d7728b39d5f496196a09689732f21ab) in parent group-v288131. [ 581.978272] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Creating folder: Instances. Parent ref: group-v288144. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 581.978272] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4a45e383-81a3-4a14-ba71-aae7b0b4faab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.991148] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Created folder: Instances in parent group-v288144. [ 581.991399] env[62914]: DEBUG oslo.service.loopingcall [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 581.993491] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 581.993491] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-637fc156-25eb-4235-9190-8af86a7a37e9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.010069] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 582.010069] env[62914]: value = "task-1352363" [ 582.010069] env[62914]: _type = "Task" [ 582.010069] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.019804] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352363, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.070483] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.477619] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b257f3d-e553-4614-86fd-6d01be0692a6 tempest-ServerDiagnosticsV248Test-1208628397 tempest-ServerDiagnosticsV248Test-1208628397-project-member] Lock "d5c90d13-41bc-4cca-abf3-5a5b13171a6f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.700s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.523125] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352363, 'name': CreateVM_Task, 'duration_secs': 0.314546} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.523325] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 582.524338] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.525897] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.526042] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 582.527031] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b430b586-2bbb-48d4-864e-3eb139cfd17d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.534992] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 582.534992] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ea7bfc-0f5c-fd0a-0697-c298cadf2ec5" [ 582.534992] env[62914]: _type = "Task" [ 582.534992] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.550445] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ea7bfc-0f5c-fd0a-0697-c298cadf2ec5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.934383] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6c61da85-db92-4980-8ef5-9d45250898aa tempest-ServersAdmin275Test-1366723344 tempest-ServersAdmin275Test-1366723344-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.935635] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.989s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.938617] env[62914]: INFO nova.compute.claims [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 583.048068] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ea7bfc-0f5c-fd0a-0697-c298cadf2ec5, 'name': SearchDatastore_Task, 'duration_secs': 0.01192} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.048395] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.048613] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 583.048834] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.048968] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.049160] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 583.049418] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1519d7e3-015e-4d6a-916b-3c21c73d36d0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.064567] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 583.064567] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 583.064567] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-576b243f-92ee-403e-bb6b-091acf98e2e8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.072377] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 583.072377] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52e6ec89-5eef-d613-d3a3-073f13226fbd" [ 583.072377] env[62914]: _type = "Task" [ 583.072377] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.083726] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52e6ec89-5eef-d613-d3a3-073f13226fbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.564148] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Acquiring lock "67554dc7-fb72-47e1-8201-f5fda035a2d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.564461] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Lock "67554dc7-fb72-47e1-8201-f5fda035a2d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.587419] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52e6ec89-5eef-d613-d3a3-073f13226fbd, 'name': SearchDatastore_Task, 'duration_secs': 0.010368} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.588448] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9e0974e-6070-4b54-aa5b-95ce2865f51c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.595043] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 583.595043] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52041fed-b385-924f-2403-d75d31037b03" [ 583.595043] env[62914]: _type = "Task" [ 583.595043] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.605845] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52041fed-b385-924f-2403-d75d31037b03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.104958] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52041fed-b385-924f-2403-d75d31037b03, 'name': SearchDatastore_Task, 'duration_secs': 0.020205} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.110309] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.110615] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67/34b0b81e-fc98-4019-92c3-fe2ce2fdfd67.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 584.111177] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-61c228de-5050-40ef-9b92-cd8e7b0b3bc9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.123940] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 584.123940] env[62914]: value = "task-1352364" [ 584.123940] env[62914]: _type = "Task" [ 584.123940] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.135722] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352364, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.498532] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98fb77b-afc5-4a57-bc85-9a3d3afea591 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.513604] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9b280b-a0b1-443d-b5c2-52bc98bd201d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.562989] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a48791-0e88-4c9e-9301-16fdb62eb637 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.577292] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb4db35b-fa34-471f-b995-f37e1ff8b65e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.600119] env[62914]: DEBUG nova.compute.provider_tree [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.640899] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352364, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479693} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.640899] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67/34b0b81e-fc98-4019-92c3-fe2ce2fdfd67.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 584.641173] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 584.641235] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1aad652c-54de-4243-86da-7f2b35e07a7d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.648109] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 584.648109] env[62914]: value = "task-1352365" [ 584.648109] env[62914]: _type = "Task" [ 584.648109] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.658399] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352365, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.101853] env[62914]: DEBUG nova.scheduler.client.report [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.166397] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352365, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059349} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.166926] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 585.167658] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5f7ca0-ae77-4c42-9e4d-eb754c00b031 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.193811] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Reconfiguring VM instance instance-00000010 to attach disk [datastore1] 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67/34b0b81e-fc98-4019-92c3-fe2ce2fdfd67.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 585.194149] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-097f7762-fab9-4c31-a4ea-b081ddbc60f6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.219159] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 585.219159] env[62914]: value = "task-1352366" [ 585.219159] env[62914]: _type = "Task" [ 585.219159] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.227659] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352366, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.611027] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.673s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.611027] env[62914]: DEBUG nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 585.614619] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.509s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.738356] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352366, 'name': ReconfigVM_Task, 'duration_secs': 0.300077} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.739386] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Reconfigured VM instance instance-00000010 to attach disk [datastore1] 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67/34b0b81e-fc98-4019-92c3-fe2ce2fdfd67.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 585.739790] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bdf2ce71-5716-4295-8075-79f6e159fcb8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.746773] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 585.746773] env[62914]: value = "task-1352367" [ 585.746773] env[62914]: _type = "Task" [ 585.746773] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.755115] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352367, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.126231] env[62914]: DEBUG nova.compute.utils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 586.126231] env[62914]: DEBUG nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 586.128053] env[62914]: DEBUG nova.network.neutron [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 586.218817] env[62914]: DEBUG nova.policy [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1edfc98c6e94f04af2b026b38d0e9db', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '600896dd961a4f1db2aa0cc9c5bf2830', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 586.261753] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352367, 'name': Rename_Task, 'duration_secs': 0.139459} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.261753] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 586.261753] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c83fb13b-5891-4cd8-aa1a-9a808b939f38 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.267833] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 586.267833] env[62914]: value = "task-1352368" [ 586.267833] env[62914]: _type = "Task" [ 586.267833] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.282842] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352368, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.640698] env[62914]: DEBUG nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 586.673993] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f5ba43-6d9a-42d5-b885-43512aac39f5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.681968] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642b6fc5-cdd1-431b-855a-e6475b0e97bb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.717317] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b56f00-3dd7-40cf-9fdd-0c414744bc8f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.728036] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7613a9e2-45c0-429a-97e7-23e8f5dc9b56 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.747217] env[62914]: DEBUG nova.compute.provider_tree [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.758397] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Acquiring lock "821366f1-f1b0-4b78-8f7e-db8fd3bb123e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.758706] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Lock "821366f1-f1b0-4b78-8f7e-db8fd3bb123e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.781910] env[62914]: DEBUG oslo_vmware.api [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352368, 'name': PowerOnVM_Task, 'duration_secs': 0.447681} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.781910] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 586.782127] env[62914]: INFO nova.compute.manager [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Took 4.92 seconds to spawn the instance on the hypervisor. [ 586.782293] env[62914]: DEBUG nova.compute.manager [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 586.783106] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296ec71d-12ae-479c-81a6-4e2071c4b7a9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.850753] env[62914]: DEBUG nova.network.neutron [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Successfully created port: ef6080b9-52f1-4d13-8cf4-b79ec17bbdf2 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 587.257763] env[62914]: DEBUG nova.scheduler.client.report [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.311329] env[62914]: INFO nova.compute.manager [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Took 33.83 seconds to build instance. [ 587.655630] env[62914]: DEBUG nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 587.690779] env[62914]: DEBUG nova.virt.hardware [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 587.690874] env[62914]: DEBUG nova.virt.hardware [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 587.691227] env[62914]: DEBUG nova.virt.hardware [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 587.691634] env[62914]: DEBUG nova.virt.hardware [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 587.691864] env[62914]: DEBUG nova.virt.hardware [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 587.692074] env[62914]: DEBUG nova.virt.hardware [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 587.692409] env[62914]: DEBUG nova.virt.hardware [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 587.692906] env[62914]: DEBUG nova.virt.hardware [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 587.693269] env[62914]: DEBUG nova.virt.hardware [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 587.693511] env[62914]: DEBUG nova.virt.hardware [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 587.694216] env[62914]: DEBUG nova.virt.hardware [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 587.695752] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524cfdf6-9240-4d8a-ab62-6a798942a8b2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.704981] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd11e70-76f0-423c-a3aa-35bdde8a6502 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.765514] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.151s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.766493] env[62914]: ERROR nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4301f15c-b22c-47fc-9616-26ee92aa1327, please check neutron logs for more information. [ 587.766493] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Traceback (most recent call last): [ 587.766493] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.766493] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] self.driver.spawn(context, instance, image_meta, [ 587.766493] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 587.766493] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.766493] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.766493] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] vm_ref = self.build_virtual_machine(instance, [ 587.766493] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.766493] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.766493] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.766887] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] for vif in network_info: [ 587.766887] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.766887] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] return self._sync_wrapper(fn, *args, **kwargs) [ 587.766887] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.766887] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] self.wait() [ 587.766887] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.766887] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] self[:] = self._gt.wait() [ 587.766887] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.766887] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] return self._exit_event.wait() [ 587.766887] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.766887] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] result = hub.switch() [ 587.766887] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.766887] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] return self.greenlet.switch() [ 587.767956] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.767956] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] result = function(*args, **kwargs) [ 587.767956] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.767956] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] return func(*args, **kwargs) [ 587.767956] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.767956] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] raise e [ 587.767956] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.767956] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] nwinfo = self.network_api.allocate_for_instance( [ 587.767956] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.767956] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] created_port_ids = self._update_ports_for_instance( [ 587.767956] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.767956] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] with excutils.save_and_reraise_exception(): [ 587.767956] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.768404] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] self.force_reraise() [ 587.768404] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.768404] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] raise self.value [ 587.768404] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.768404] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] updated_port = self._update_port( [ 587.768404] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.768404] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] _ensure_no_port_binding_failure(port) [ 587.768404] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.768404] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] raise exception.PortBindingFailed(port_id=port['id']) [ 587.768404] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] nova.exception.PortBindingFailed: Binding failed for port 4301f15c-b22c-47fc-9616-26ee92aa1327, please check neutron logs for more information. [ 587.768404] env[62914]: ERROR nova.compute.manager [instance: 889af877-29c4-4c06-9995-49fb15b21a55] [ 587.768741] env[62914]: DEBUG nova.compute.utils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Binding failed for port 4301f15c-b22c-47fc-9616-26ee92aa1327, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 587.768741] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.554s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.772010] env[62914]: DEBUG nova.objects.instance [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Lazy-loading 'resources' on Instance uuid 3786cd93-1a50-4997-8894-840afa1c8417 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 587.772010] env[62914]: DEBUG nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Build of instance 889af877-29c4-4c06-9995-49fb15b21a55 was re-scheduled: Binding failed for port 4301f15c-b22c-47fc-9616-26ee92aa1327, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 587.772010] env[62914]: DEBUG nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 587.772694] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Acquiring lock "refresh_cache-889af877-29c4-4c06-9995-49fb15b21a55" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.772920] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Acquired lock "refresh_cache-889af877-29c4-4c06-9995-49fb15b21a55" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.773307] env[62914]: DEBUG nova.network.neutron [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 587.814136] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9034098-6844-423a-a83a-79e582be1287 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "34b0b81e-fc98-4019-92c3-fe2ce2fdfd67" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.280s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.117905] env[62914]: ERROR nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ef6080b9-52f1-4d13-8cf4-b79ec17bbdf2, please check neutron logs for more information. [ 588.117905] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 588.117905] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.117905] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 588.117905] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.117905] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 588.117905] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.117905] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 588.117905] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.117905] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 588.117905] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.117905] env[62914]: ERROR nova.compute.manager raise self.value [ 588.117905] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.117905] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 588.117905] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.117905] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 588.118613] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.118613] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 588.118613] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ef6080b9-52f1-4d13-8cf4-b79ec17bbdf2, please check neutron logs for more information. [ 588.118613] env[62914]: ERROR nova.compute.manager [ 588.118613] env[62914]: Traceback (most recent call last): [ 588.118613] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 588.118613] env[62914]: listener.cb(fileno) [ 588.118613] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.118613] env[62914]: result = function(*args, **kwargs) [ 588.118613] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 588.118613] env[62914]: return func(*args, **kwargs) [ 588.118613] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.118613] env[62914]: raise e [ 588.118613] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.118613] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 588.118613] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.118613] env[62914]: created_port_ids = self._update_ports_for_instance( [ 588.118613] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.118613] env[62914]: with excutils.save_and_reraise_exception(): [ 588.118613] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.118613] env[62914]: self.force_reraise() [ 588.118613] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.118613] env[62914]: raise self.value [ 588.118613] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.118613] env[62914]: updated_port = self._update_port( [ 588.118613] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.118613] env[62914]: _ensure_no_port_binding_failure(port) [ 588.118613] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.118613] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 588.119506] env[62914]: nova.exception.PortBindingFailed: Binding failed for port ef6080b9-52f1-4d13-8cf4-b79ec17bbdf2, please check neutron logs for more information. [ 588.119506] env[62914]: Removing descriptor: 15 [ 588.119506] env[62914]: ERROR nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ef6080b9-52f1-4d13-8cf4-b79ec17bbdf2, please check neutron logs for more information. [ 588.119506] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] Traceback (most recent call last): [ 588.119506] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 588.119506] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] yield resources [ 588.119506] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.119506] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] self.driver.spawn(context, instance, image_meta, [ 588.119506] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 588.119506] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.119506] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.119506] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] vm_ref = self.build_virtual_machine(instance, [ 588.119985] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.119985] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.119985] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.119985] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] for vif in network_info: [ 588.119985] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.119985] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] return self._sync_wrapper(fn, *args, **kwargs) [ 588.119985] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.119985] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] self.wait() [ 588.119985] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.119985] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] self[:] = self._gt.wait() [ 588.119985] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.119985] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] return self._exit_event.wait() [ 588.119985] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.120421] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] result = hub.switch() [ 588.120421] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.120421] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] return self.greenlet.switch() [ 588.120421] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.120421] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] result = function(*args, **kwargs) [ 588.120421] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 588.120421] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] return func(*args, **kwargs) [ 588.120421] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.120421] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] raise e [ 588.120421] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.120421] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] nwinfo = self.network_api.allocate_for_instance( [ 588.120421] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.120421] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] created_port_ids = self._update_ports_for_instance( [ 588.120811] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.120811] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] with excutils.save_and_reraise_exception(): [ 588.120811] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.120811] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] self.force_reraise() [ 588.120811] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.120811] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] raise self.value [ 588.120811] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.120811] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] updated_port = self._update_port( [ 588.120811] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.120811] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] _ensure_no_port_binding_failure(port) [ 588.120811] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.120811] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] raise exception.PortBindingFailed(port_id=port['id']) [ 588.122082] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] nova.exception.PortBindingFailed: Binding failed for port ef6080b9-52f1-4d13-8cf4-b79ec17bbdf2, please check neutron logs for more information. [ 588.122082] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] [ 588.122082] env[62914]: INFO nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Terminating instance [ 588.125079] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Acquiring lock "refresh_cache-723105bc-6810-4759-8240-af0a128475e4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.125079] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Acquired lock "refresh_cache-723105bc-6810-4759-8240-af0a128475e4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.125079] env[62914]: DEBUG nova.network.neutron [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.152820] env[62914]: DEBUG nova.compute.manager [req-0b8e9b32-c35d-4f62-92f8-421a14af0c6c req-efe9191c-5a50-46eb-a335-8cb718bba204 service nova] [instance: 723105bc-6810-4759-8240-af0a128475e4] Received event network-changed-ef6080b9-52f1-4d13-8cf4-b79ec17bbdf2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 588.152820] env[62914]: DEBUG nova.compute.manager [req-0b8e9b32-c35d-4f62-92f8-421a14af0c6c req-efe9191c-5a50-46eb-a335-8cb718bba204 service nova] [instance: 723105bc-6810-4759-8240-af0a128475e4] Refreshing instance network info cache due to event network-changed-ef6080b9-52f1-4d13-8cf4-b79ec17bbdf2. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 588.152820] env[62914]: DEBUG oslo_concurrency.lockutils [req-0b8e9b32-c35d-4f62-92f8-421a14af0c6c req-efe9191c-5a50-46eb-a335-8cb718bba204 service nova] Acquiring lock "refresh_cache-723105bc-6810-4759-8240-af0a128475e4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.310617] env[62914]: DEBUG nova.network.neutron [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.315441] env[62914]: DEBUG nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 588.381514] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Acquiring lock "9edfad51-10a5-4e0c-a337-2de909c361cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.381755] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Lock "9edfad51-10a5-4e0c-a337-2de909c361cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.440391] env[62914]: DEBUG nova.network.neutron [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.646474] env[62914]: DEBUG nova.network.neutron [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.774704] env[62914]: DEBUG nova.network.neutron [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.809285] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fafdf53-ca1f-4592-8124-9a5776ddb8a3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.818975] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ad1163-1fca-4356-a96f-a4cbc64afdfb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.867391] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.868759] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea73b5a9-850a-4646-8a08-7f1907901e7a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.875992] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71edcaa-06d1-4340-a0e8-d391b4fa8c69 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.894946] env[62914]: DEBUG nova.compute.provider_tree [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.942714] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Releasing lock "refresh_cache-889af877-29c4-4c06-9995-49fb15b21a55" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.942980] env[62914]: DEBUG nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 588.943188] env[62914]: DEBUG nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 588.943364] env[62914]: DEBUG nova.network.neutron [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 588.972461] env[62914]: DEBUG nova.network.neutron [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.279675] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Releasing lock "refresh_cache-723105bc-6810-4759-8240-af0a128475e4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.280058] env[62914]: DEBUG nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.280571] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 589.280571] env[62914]: DEBUG oslo_concurrency.lockutils [req-0b8e9b32-c35d-4f62-92f8-421a14af0c6c req-efe9191c-5a50-46eb-a335-8cb718bba204 service nova] Acquired lock "refresh_cache-723105bc-6810-4759-8240-af0a128475e4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.280791] env[62914]: DEBUG nova.network.neutron [req-0b8e9b32-c35d-4f62-92f8-421a14af0c6c req-efe9191c-5a50-46eb-a335-8cb718bba204 service nova] [instance: 723105bc-6810-4759-8240-af0a128475e4] Refreshing network info cache for port ef6080b9-52f1-4d13-8cf4-b79ec17bbdf2 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 589.283060] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f4c8c01-d985-4b3a-af62-6ec33655f944 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.292865] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9e147b-0731-4538-ae3f-931e52fb6f96 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.319651] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 723105bc-6810-4759-8240-af0a128475e4 could not be found. [ 589.319651] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 589.319651] env[62914]: INFO nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 589.319651] env[62914]: DEBUG oslo.service.loopingcall [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 589.319651] env[62914]: DEBUG nova.compute.manager [-] [instance: 723105bc-6810-4759-8240-af0a128475e4] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.319651] env[62914]: DEBUG nova.network.neutron [-] [instance: 723105bc-6810-4759-8240-af0a128475e4] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 589.352732] env[62914]: DEBUG nova.network.neutron [-] [instance: 723105bc-6810-4759-8240-af0a128475e4] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.385252] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Acquiring lock "e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.385533] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Lock "e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.400672] env[62914]: DEBUG nova.scheduler.client.report [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.476610] env[62914]: DEBUG nova.network.neutron [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.814565] env[62914]: DEBUG nova.network.neutron [req-0b8e9b32-c35d-4f62-92f8-421a14af0c6c req-efe9191c-5a50-46eb-a335-8cb718bba204 service nova] [instance: 723105bc-6810-4759-8240-af0a128475e4] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.855146] env[62914]: DEBUG nova.network.neutron [-] [instance: 723105bc-6810-4759-8240-af0a128475e4] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.909979] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.142s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.913318] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.984s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.914394] env[62914]: INFO nova.compute.claims [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 589.938429] env[62914]: INFO nova.scheduler.client.report [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Deleted allocations for instance 3786cd93-1a50-4997-8894-840afa1c8417 [ 589.980680] env[62914]: INFO nova.compute.manager [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] [instance: 889af877-29c4-4c06-9995-49fb15b21a55] Took 1.04 seconds to deallocate network for instance. [ 589.991063] env[62914]: DEBUG nova.network.neutron [req-0b8e9b32-c35d-4f62-92f8-421a14af0c6c req-efe9191c-5a50-46eb-a335-8cb718bba204 service nova] [instance: 723105bc-6810-4759-8240-af0a128475e4] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.311495] env[62914]: DEBUG nova.compute.manager [req-d8e2a832-bc12-433e-a566-31ed32bd327c req-ec949b4d-49f8-428d-a882-08e056194340 service nova] [instance: 723105bc-6810-4759-8240-af0a128475e4] Received event network-vif-deleted-ef6080b9-52f1-4d13-8cf4-b79ec17bbdf2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 590.357602] env[62914]: INFO nova.compute.manager [-] [instance: 723105bc-6810-4759-8240-af0a128475e4] Took 1.04 seconds to deallocate network for instance. [ 590.361822] env[62914]: DEBUG nova.compute.claims [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 590.361996] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.453025] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0c44f8d-3dbd-478a-b680-9c25d2ab0ade tempest-ServersAdmin275Test-845109781 tempest-ServersAdmin275Test-845109781-project-member] Lock "3786cd93-1a50-4997-8894-840afa1c8417" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.120s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.493685] env[62914]: DEBUG oslo_concurrency.lockutils [req-0b8e9b32-c35d-4f62-92f8-421a14af0c6c req-efe9191c-5a50-46eb-a335-8cb718bba204 service nova] Releasing lock "refresh_cache-723105bc-6810-4759-8240-af0a128475e4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.710667] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "8399d066-43bc-4766-97c8-89d74f3283f0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.710914] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "8399d066-43bc-4766-97c8-89d74f3283f0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.021009] env[62914]: INFO nova.scheduler.client.report [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Deleted allocations for instance 889af877-29c4-4c06-9995-49fb15b21a55 [ 591.495732] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d30d6c-cf5f-49ee-889b-943d676207b0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.505589] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80ab272-0cb0-432b-b546-f150db7d2f37 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.545519] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35062e8c-ddf0-4b06-a661-b7e6b1b96cea tempest-ServersWithSpecificFlavorTestJSON-1744821619 tempest-ServersWithSpecificFlavorTestJSON-1744821619-project-member] Lock "889af877-29c4-4c06-9995-49fb15b21a55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.672s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.548862] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5a7ea5-4f6f-4454-b96c-e13dd7fb4f1a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.561899] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf9ee261-05a1-4951-a5ee-7c00d0c2aa6c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.578816] env[62914]: DEBUG nova.compute.provider_tree [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.055940] env[62914]: DEBUG nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 592.084311] env[62914]: DEBUG nova.scheduler.client.report [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.586488] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.592702] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.680s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.593098] env[62914]: DEBUG nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 592.596155] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.721s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.100928] env[62914]: DEBUG nova.compute.utils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 593.107631] env[62914]: DEBUG nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 593.107935] env[62914]: DEBUG nova.network.neutron [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 593.182196] env[62914]: DEBUG nova.policy [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e502b585cc3546e191a3228cb154322c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '95ab27c882c7450dbfeeac909d5be4b1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 593.590095] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a435a369-61b7-4557-8169-fb1b6ff6f053 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.602929] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-164d4b1b-b10c-45b2-abbe-1b67fcc5d18a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.607045] env[62914]: DEBUG nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 593.640459] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-111b8fb9-679b-47d5-b896-4a93a867caff {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.653407] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db46477-3be0-47b1-85b9-669aaeac685e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.672186] env[62914]: DEBUG nova.compute.provider_tree [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.747223] env[62914]: DEBUG nova.network.neutron [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Successfully created port: b81867b2-0504-426f-9262-65caf32e66fd {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 594.175286] env[62914]: DEBUG nova.scheduler.client.report [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.647896] env[62914]: DEBUG nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 594.685581] env[62914]: DEBUG nova.virt.hardware [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 594.685874] env[62914]: DEBUG nova.virt.hardware [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 594.685975] env[62914]: DEBUG nova.virt.hardware [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 594.686165] env[62914]: DEBUG nova.virt.hardware [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 594.686315] env[62914]: DEBUG nova.virt.hardware [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 594.686453] env[62914]: DEBUG nova.virt.hardware [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 594.686694] env[62914]: DEBUG nova.virt.hardware [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 594.686810] env[62914]: DEBUG nova.virt.hardware [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 594.686970] env[62914]: DEBUG nova.virt.hardware [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 594.687146] env[62914]: DEBUG nova.virt.hardware [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 594.687315] env[62914]: DEBUG nova.virt.hardware [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 594.688697] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.092s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.694212] env[62914]: ERROR nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8f6beb45-a893-4500-bfb4-d92c2f46c170, please check neutron logs for more information. [ 594.694212] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Traceback (most recent call last): [ 594.694212] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 594.694212] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] self.driver.spawn(context, instance, image_meta, [ 594.694212] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 594.694212] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.694212] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.694212] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] vm_ref = self.build_virtual_machine(instance, [ 594.694212] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.694212] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.694212] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.694484] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] for vif in network_info: [ 594.694484] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.694484] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] return self._sync_wrapper(fn, *args, **kwargs) [ 594.694484] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.694484] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] self.wait() [ 594.694484] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.694484] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] self[:] = self._gt.wait() [ 594.694484] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.694484] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] return self._exit_event.wait() [ 594.694484] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.694484] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] result = hub.switch() [ 594.694484] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.694484] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] return self.greenlet.switch() [ 594.694767] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.694767] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] result = function(*args, **kwargs) [ 594.694767] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 594.694767] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] return func(*args, **kwargs) [ 594.694767] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.694767] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] raise e [ 594.694767] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.694767] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] nwinfo = self.network_api.allocate_for_instance( [ 594.694767] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.694767] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] created_port_ids = self._update_ports_for_instance( [ 594.694767] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.694767] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] with excutils.save_and_reraise_exception(): [ 594.694767] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.695072] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] self.force_reraise() [ 594.695072] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.695072] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] raise self.value [ 594.695072] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.695072] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] updated_port = self._update_port( [ 594.695072] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.695072] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] _ensure_no_port_binding_failure(port) [ 594.695072] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.695072] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] raise exception.PortBindingFailed(port_id=port['id']) [ 594.695072] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] nova.exception.PortBindingFailed: Binding failed for port 8f6beb45-a893-4500-bfb4-d92c2f46c170, please check neutron logs for more information. [ 594.695072] env[62914]: ERROR nova.compute.manager [instance: 4981a04f-98e6-474b-850a-634d977d9e42] [ 594.695580] env[62914]: DEBUG nova.compute.utils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Binding failed for port 8f6beb45-a893-4500-bfb4-d92c2f46c170, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 594.696773] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c546d09-43bc-41ff-aa0c-aa4c7e15cad3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.706895] env[62914]: DEBUG nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Build of instance 4981a04f-98e6-474b-850a-634d977d9e42 was re-scheduled: Binding failed for port 8f6beb45-a893-4500-bfb4-d92c2f46c170, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 594.706895] env[62914]: DEBUG nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 594.707189] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Acquiring lock "refresh_cache-4981a04f-98e6-474b-850a-634d977d9e42" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.707407] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Acquired lock "refresh_cache-4981a04f-98e6-474b-850a-634d977d9e42" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.707651] env[62914]: DEBUG nova.network.neutron [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 594.708904] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.082s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.726933] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-750f08be-9be8-469f-95f8-d1f6eef2ad19 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.407840] env[62914]: DEBUG nova.network.neutron [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.632851] env[62914]: DEBUG nova.network.neutron [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.752968] env[62914]: DEBUG nova.compute.manager [req-f38908fe-edbc-4e51-b8c1-f6e2b3cd6756 req-bd43489a-d305-46b0-8a67-04d58dd3db33 service nova] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Received event network-changed-b81867b2-0504-426f-9262-65caf32e66fd {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 595.753355] env[62914]: DEBUG nova.compute.manager [req-f38908fe-edbc-4e51-b8c1-f6e2b3cd6756 req-bd43489a-d305-46b0-8a67-04d58dd3db33 service nova] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Refreshing instance network info cache due to event network-changed-b81867b2-0504-426f-9262-65caf32e66fd. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 595.753704] env[62914]: DEBUG oslo_concurrency.lockutils [req-f38908fe-edbc-4e51-b8c1-f6e2b3cd6756 req-bd43489a-d305-46b0-8a67-04d58dd3db33 service nova] Acquiring lock "refresh_cache-bfcba422-5ed0-4bfd-83e9-cdde324d899f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.753947] env[62914]: DEBUG oslo_concurrency.lockutils [req-f38908fe-edbc-4e51-b8c1-f6e2b3cd6756 req-bd43489a-d305-46b0-8a67-04d58dd3db33 service nova] Acquired lock "refresh_cache-bfcba422-5ed0-4bfd-83e9-cdde324d899f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.754232] env[62914]: DEBUG nova.network.neutron [req-f38908fe-edbc-4e51-b8c1-f6e2b3cd6756 req-bd43489a-d305-46b0-8a67-04d58dd3db33 service nova] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Refreshing network info cache for port b81867b2-0504-426f-9262-65caf32e66fd {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 595.764868] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b03eed1-3046-42e3-8e52-a67f6292185a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.777020] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f712c89-31b4-45d7-9e04-0b46c8355f6e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.809791] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34591b49-1594-4835-9e2e-9c8c59eb3a6a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.817792] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c7fc9f7-19eb-4aed-9bc3-53c45107d59d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.832742] env[62914]: DEBUG nova.compute.provider_tree [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.957108] env[62914]: ERROR nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b81867b2-0504-426f-9262-65caf32e66fd, please check neutron logs for more information. [ 595.957108] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 595.957108] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.957108] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 595.957108] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.957108] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 595.957108] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.957108] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 595.957108] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.957108] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 595.957108] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.957108] env[62914]: ERROR nova.compute.manager raise self.value [ 595.957108] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.957108] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 595.957108] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.957108] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 595.957468] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.957468] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 595.957468] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b81867b2-0504-426f-9262-65caf32e66fd, please check neutron logs for more information. [ 595.957468] env[62914]: ERROR nova.compute.manager [ 595.957468] env[62914]: Traceback (most recent call last): [ 595.957468] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 595.957468] env[62914]: listener.cb(fileno) [ 595.957468] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.957468] env[62914]: result = function(*args, **kwargs) [ 595.957468] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 595.957468] env[62914]: return func(*args, **kwargs) [ 595.957468] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.957468] env[62914]: raise e [ 595.957468] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.957468] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 595.957468] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.957468] env[62914]: created_port_ids = self._update_ports_for_instance( [ 595.957468] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.957468] env[62914]: with excutils.save_and_reraise_exception(): [ 595.957468] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.957468] env[62914]: self.force_reraise() [ 595.957468] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.957468] env[62914]: raise self.value [ 595.957468] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.957468] env[62914]: updated_port = self._update_port( [ 595.957468] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.957468] env[62914]: _ensure_no_port_binding_failure(port) [ 595.957468] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.957468] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 595.958120] env[62914]: nova.exception.PortBindingFailed: Binding failed for port b81867b2-0504-426f-9262-65caf32e66fd, please check neutron logs for more information. [ 595.958120] env[62914]: Removing descriptor: 15 [ 595.958120] env[62914]: ERROR nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b81867b2-0504-426f-9262-65caf32e66fd, please check neutron logs for more information. [ 595.958120] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Traceback (most recent call last): [ 595.958120] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 595.958120] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] yield resources [ 595.958120] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 595.958120] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] self.driver.spawn(context, instance, image_meta, [ 595.958120] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 595.958120] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.958120] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.958120] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] vm_ref = self.build_virtual_machine(instance, [ 595.958385] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.958385] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.958385] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.958385] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] for vif in network_info: [ 595.958385] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.958385] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] return self._sync_wrapper(fn, *args, **kwargs) [ 595.958385] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.958385] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] self.wait() [ 595.958385] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.958385] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] self[:] = self._gt.wait() [ 595.958385] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.958385] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] return self._exit_event.wait() [ 595.958385] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.958641] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] result = hub.switch() [ 595.958641] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.958641] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] return self.greenlet.switch() [ 595.958641] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.958641] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] result = function(*args, **kwargs) [ 595.958641] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 595.958641] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] return func(*args, **kwargs) [ 595.958641] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.958641] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] raise e [ 595.958641] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.958641] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] nwinfo = self.network_api.allocate_for_instance( [ 595.958641] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.958641] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] created_port_ids = self._update_ports_for_instance( [ 595.958974] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.958974] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] with excutils.save_and_reraise_exception(): [ 595.958974] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.958974] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] self.force_reraise() [ 595.958974] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.958974] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] raise self.value [ 595.958974] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.958974] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] updated_port = self._update_port( [ 595.958974] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.958974] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] _ensure_no_port_binding_failure(port) [ 595.958974] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.958974] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] raise exception.PortBindingFailed(port_id=port['id']) [ 595.959324] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] nova.exception.PortBindingFailed: Binding failed for port b81867b2-0504-426f-9262-65caf32e66fd, please check neutron logs for more information. [ 595.959324] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] [ 595.959324] env[62914]: INFO nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Terminating instance [ 595.960970] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Acquiring lock "refresh_cache-bfcba422-5ed0-4bfd-83e9-cdde324d899f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.142189] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Releasing lock "refresh_cache-4981a04f-98e6-474b-850a-634d977d9e42" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.142794] env[62914]: DEBUG nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 596.143059] env[62914]: DEBUG nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.143461] env[62914]: DEBUG nova.network.neutron [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 596.175163] env[62914]: DEBUG nova.network.neutron [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.285589] env[62914]: DEBUG nova.network.neutron [req-f38908fe-edbc-4e51-b8c1-f6e2b3cd6756 req-bd43489a-d305-46b0-8a67-04d58dd3db33 service nova] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.335913] env[62914]: DEBUG nova.scheduler.client.report [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.412112] env[62914]: DEBUG nova.network.neutron [req-f38908fe-edbc-4e51-b8c1-f6e2b3cd6756 req-bd43489a-d305-46b0-8a67-04d58dd3db33 service nova] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.677845] env[62914]: DEBUG nova.network.neutron [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.844203] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.135s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.844894] env[62914]: ERROR nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 34b9a0ca-528b-46fc-8631-87bc3661e69b, please check neutron logs for more information. [ 596.844894] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Traceback (most recent call last): [ 596.844894] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.844894] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] self.driver.spawn(context, instance, image_meta, [ 596.844894] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 596.844894] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.844894] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.844894] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] vm_ref = self.build_virtual_machine(instance, [ 596.844894] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.844894] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.844894] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.845268] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] for vif in network_info: [ 596.845268] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.845268] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] return self._sync_wrapper(fn, *args, **kwargs) [ 596.845268] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.845268] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] self.wait() [ 596.845268] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.845268] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] self[:] = self._gt.wait() [ 596.845268] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.845268] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] return self._exit_event.wait() [ 596.845268] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.845268] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] result = hub.switch() [ 596.845268] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.845268] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] return self.greenlet.switch() [ 596.845580] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.845580] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] result = function(*args, **kwargs) [ 596.845580] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 596.845580] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] return func(*args, **kwargs) [ 596.845580] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.845580] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] raise e [ 596.845580] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.845580] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] nwinfo = self.network_api.allocate_for_instance( [ 596.845580] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.845580] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] created_port_ids = self._update_ports_for_instance( [ 596.845580] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.845580] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] with excutils.save_and_reraise_exception(): [ 596.845580] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.845841] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] self.force_reraise() [ 596.845841] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.845841] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] raise self.value [ 596.845841] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.845841] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] updated_port = self._update_port( [ 596.845841] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.845841] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] _ensure_no_port_binding_failure(port) [ 596.845841] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.845841] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] raise exception.PortBindingFailed(port_id=port['id']) [ 596.845841] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] nova.exception.PortBindingFailed: Binding failed for port 34b9a0ca-528b-46fc-8631-87bc3661e69b, please check neutron logs for more information. [ 596.845841] env[62914]: ERROR nova.compute.manager [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] [ 596.846106] env[62914]: DEBUG nova.compute.utils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Binding failed for port 34b9a0ca-528b-46fc-8631-87bc3661e69b, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 596.848314] env[62914]: DEBUG nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Build of instance f1cee1cb-aceb-47e8-b2da-b94563f922e2 was re-scheduled: Binding failed for port 34b9a0ca-528b-46fc-8631-87bc3661e69b, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 596.848368] env[62914]: DEBUG nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 596.848608] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Acquiring lock "refresh_cache-f1cee1cb-aceb-47e8-b2da-b94563f922e2" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.848686] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Acquired lock "refresh_cache-f1cee1cb-aceb-47e8-b2da-b94563f922e2" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.848836] env[62914]: DEBUG nova.network.neutron [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.850689] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.548s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.852768] env[62914]: INFO nova.compute.claims [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.915088] env[62914]: DEBUG oslo_concurrency.lockutils [req-f38908fe-edbc-4e51-b8c1-f6e2b3cd6756 req-bd43489a-d305-46b0-8a67-04d58dd3db33 service nova] Releasing lock "refresh_cache-bfcba422-5ed0-4bfd-83e9-cdde324d899f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.915579] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Acquired lock "refresh_cache-bfcba422-5ed0-4bfd-83e9-cdde324d899f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.915779] env[62914]: DEBUG nova.network.neutron [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 597.186019] env[62914]: INFO nova.compute.manager [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] [instance: 4981a04f-98e6-474b-850a-634d977d9e42] Took 1.04 seconds to deallocate network for instance. [ 597.435752] env[62914]: DEBUG nova.network.neutron [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.548570] env[62914]: DEBUG nova.network.neutron [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.666705] env[62914]: DEBUG nova.network.neutron [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.692357] env[62914]: DEBUG nova.network.neutron [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.086298] env[62914]: DEBUG nova.compute.manager [req-3521091b-b0c5-4df0-9914-e525b54b526c req-60923e2f-6998-4516-8d45-42ee998be937 service nova] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Received event network-vif-deleted-b81867b2-0504-426f-9262-65caf32e66fd {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 598.171249] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Releasing lock "refresh_cache-f1cee1cb-aceb-47e8-b2da-b94563f922e2" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.171249] env[62914]: DEBUG nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 598.171249] env[62914]: DEBUG nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.171249] env[62914]: DEBUG nova.network.neutron [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 598.197223] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Releasing lock "refresh_cache-bfcba422-5ed0-4bfd-83e9-cdde324d899f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.197635] env[62914]: DEBUG nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 598.197903] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 598.200903] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c365d875-a69c-42a1-8ace-2a49e0da1d8f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.207579] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c04431-a62f-4545-a216-d572d76ef8bc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.221385] env[62914]: INFO nova.scheduler.client.report [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Deleted allocations for instance 4981a04f-98e6-474b-850a-634d977d9e42 [ 598.227901] env[62914]: DEBUG nova.network.neutron [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.244275] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bfcba422-5ed0-4bfd-83e9-cdde324d899f could not be found. [ 598.244492] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 598.244664] env[62914]: INFO nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 598.244909] env[62914]: DEBUG oslo.service.loopingcall [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 598.245152] env[62914]: DEBUG nova.compute.manager [-] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.245257] env[62914]: DEBUG nova.network.neutron [-] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 598.277890] env[62914]: DEBUG nova.network.neutron [-] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.381925] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c93bde-cf32-4e05-bee3-3ea7acbf2162 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.390149] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e9bd183-742b-4d36-a571-3b9d58e1b767 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.421947] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac5d07c-d83f-473c-8719-2858f614a549 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.433172] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb0ab144-f6cf-4db5-8b9f-e76d7baff68a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.450296] env[62914]: DEBUG nova.compute.provider_tree [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.725980] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "4c1016c4-2da8-4f4e-9497-4d680fb74535" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.726700] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "4c1016c4-2da8-4f4e-9497-4d680fb74535" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.733076] env[62914]: DEBUG nova.network.neutron [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.734955] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b6c28626-0568-448b-acb5-3c782e564578 tempest-ServerActionsV293TestJSON-1140230643 tempest-ServerActionsV293TestJSON-1140230643-project-member] Lock "4981a04f-98e6-474b-850a-634d977d9e42" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.679s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.766069] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "41218b72-0b79-496a-855f-57a972dabd0b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.766448] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "41218b72-0b79-496a-855f-57a972dabd0b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.781513] env[62914]: DEBUG nova.network.neutron [-] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.956294] env[62914]: DEBUG nova.scheduler.client.report [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.236110] env[62914]: INFO nova.compute.manager [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] [instance: f1cee1cb-aceb-47e8-b2da-b94563f922e2] Took 1.07 seconds to deallocate network for instance. [ 599.241978] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 599.284470] env[62914]: INFO nova.compute.manager [-] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Took 1.04 seconds to deallocate network for instance. [ 599.288171] env[62914]: DEBUG nova.compute.claims [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 599.288171] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.463322] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.612s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.463696] env[62914]: DEBUG nova.compute.manager [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 599.469494] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.663s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.470787] env[62914]: INFO nova.compute.claims [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.775440] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.976422] env[62914]: DEBUG nova.compute.utils [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 599.981065] env[62914]: DEBUG nova.compute.manager [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Not allocating networking since 'none' was specified. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 600.228648] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Acquiring lock "d9057a9e-988f-44ee-9339-ca9a1429378e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.228837] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Lock "d9057a9e-988f-44ee-9339-ca9a1429378e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.285356] env[62914]: INFO nova.scheduler.client.report [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Deleted allocations for instance f1cee1cb-aceb-47e8-b2da-b94563f922e2 [ 600.482901] env[62914]: DEBUG nova.compute.manager [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 600.796742] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ece450a1-5855-413a-bb46-02d7636ffae4 tempest-ServersAdminNegativeTestJSON-877198869 tempest-ServersAdminNegativeTestJSON-877198869-project-member] Lock "f1cee1cb-aceb-47e8-b2da-b94563f922e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.705s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.843229] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Acquiring lock "220578cb-17b1-4b46-8c4a-f676c04fc5e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.843514] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Lock "220578cb-17b1-4b46-8c4a-f676c04fc5e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.026097] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ff0d48-edaf-4c2a-856d-583815a941eb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.034464] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f23515-ff87-4ab5-bb08-4a7ce64801f6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.077350] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698159c8-8485-4c02-9668-14d4655f5886 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.085889] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e7753f-ee5f-4d74-bba5-ce4bc1ef8c0c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.101916] env[62914]: DEBUG nova.compute.provider_tree [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.140689] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Acquiring lock "d73f5ea8-529a-44e3-8239-5aaa4534d1cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.140920] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Lock "d73f5ea8-529a-44e3-8239-5aaa4534d1cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.300999] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 601.497911] env[62914]: DEBUG nova.compute.manager [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 601.543818] env[62914]: DEBUG nova.virt.hardware [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 601.544155] env[62914]: DEBUG nova.virt.hardware [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 601.544230] env[62914]: DEBUG nova.virt.hardware [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 601.544847] env[62914]: DEBUG nova.virt.hardware [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 601.544847] env[62914]: DEBUG nova.virt.hardware [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 601.544847] env[62914]: DEBUG nova.virt.hardware [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 601.545346] env[62914]: DEBUG nova.virt.hardware [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 601.545346] env[62914]: DEBUG nova.virt.hardware [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 601.546104] env[62914]: DEBUG nova.virt.hardware [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 601.546295] env[62914]: DEBUG nova.virt.hardware [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 601.546506] env[62914]: DEBUG nova.virt.hardware [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.547396] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc65b6c6-8312-4472-a2ca-c93ae2fdf2a2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.555882] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faceee3b-3d7a-4db2-87d2-1e5a12465719 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.570817] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Instance VIF info [] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 601.576706] env[62914]: DEBUG oslo.service.loopingcall [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 601.576995] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 601.577867] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-78b9aebe-9f32-47f2-b50f-28634b71b997 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.596047] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 601.596047] env[62914]: value = "task-1352370" [ 601.596047] env[62914]: _type = "Task" [ 601.596047] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.604512] env[62914]: DEBUG nova.scheduler.client.report [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.608112] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352370, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.840765] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.106035] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352370, 'name': CreateVM_Task, 'duration_secs': 0.311437} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.106242] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 602.106686] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.106843] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.107177] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 602.107418] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5eeee233-857c-4af1-9aac-f1a37eb40db5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.109906] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.641s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.110388] env[62914]: DEBUG nova.compute.manager [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.114172] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.900s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.121600] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 602.121600] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a6217a-6e26-44ab-3630-9105319ad439" [ 602.121600] env[62914]: _type = "Task" [ 602.121600] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.133423] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a6217a-6e26-44ab-3630-9105319ad439, 'name': SearchDatastore_Task, 'duration_secs': 0.009303} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.134036] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.134036] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 602.134533] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.134685] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.134873] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 602.135134] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a07f1f2-1390-4a9d-ac57-b0d81fe08b3c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.146020] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 602.146020] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 602.146020] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-666ebce2-19d2-4e59-8ab2-f0c7e43979e7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.149926] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 602.149926] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]526c47fd-e212-a66e-5b97-594fbc8b2c90" [ 602.149926] env[62914]: _type = "Task" [ 602.149926] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.159276] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]526c47fd-e212-a66e-5b97-594fbc8b2c90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.623389] env[62914]: DEBUG nova.compute.utils [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.627774] env[62914]: DEBUG nova.compute.manager [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Not allocating networking since 'none' was specified. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 602.671872] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]526c47fd-e212-a66e-5b97-594fbc8b2c90, 'name': SearchDatastore_Task, 'duration_secs': 0.019358} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.675213] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1cdf7ea-844d-41ae-b72b-86312924f132 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.680571] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 602.680571] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]521100ce-9bce-d003-b8f8-51b3d1bc6428" [ 602.680571] env[62914]: _type = "Task" [ 602.680571] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.689253] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]521100ce-9bce-d003-b8f8-51b3d1bc6428, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.116865] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-777c291c-1700-4e8c-9b5e-eb7a349a5314 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.125013] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47be40fd-7dc7-48cd-9322-dbcb563d4b5e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.128842] env[62914]: DEBUG nova.compute.manager [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 603.164222] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824aecd1-d8fd-4de2-b393-9498af60a49b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.174525] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c45e2f2-dee1-4e98-b308-b97bdccdf5b5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.189069] env[62914]: DEBUG nova.compute.provider_tree [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.202675] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]521100ce-9bce-d003-b8f8-51b3d1bc6428, 'name': SearchDatastore_Task, 'duration_secs': 0.009279} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.203869] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.203869] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] f3509973-013e-45ce-87f9-357e782f26d7/f3509973-013e-45ce-87f9-357e782f26d7.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 603.203997] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-13b82e2e-2ea9-4cb7-bc39-cf8f76bb95e8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.210614] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 603.210614] env[62914]: value = "task-1352371" [ 603.210614] env[62914]: _type = "Task" [ 603.210614] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.219050] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352371, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.698442] env[62914]: DEBUG nova.scheduler.client.report [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.723754] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352371, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502701} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.723754] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] f3509973-013e-45ce-87f9-357e782f26d7/f3509973-013e-45ce-87f9-357e782f26d7.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 603.723754] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 603.723879] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e01d073f-51ad-468b-a008-f893708fe76c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.730229] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 603.730229] env[62914]: value = "task-1352372" [ 603.730229] env[62914]: _type = "Task" [ 603.730229] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.740474] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352372, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.132037] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquiring lock "44339b45-5183-4882-aaec-4070adc3c3dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.132037] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Lock "44339b45-5183-4882-aaec-4070adc3c3dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.137724] env[62914]: DEBUG nova.compute.manager [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.170733] env[62914]: DEBUG nova.virt.hardware [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.170977] env[62914]: DEBUG nova.virt.hardware [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.171166] env[62914]: DEBUG nova.virt.hardware [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.171386] env[62914]: DEBUG nova.virt.hardware [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.171492] env[62914]: DEBUG nova.virt.hardware [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.171633] env[62914]: DEBUG nova.virt.hardware [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.172247] env[62914]: DEBUG nova.virt.hardware [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.172247] env[62914]: DEBUG nova.virt.hardware [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.172247] env[62914]: DEBUG nova.virt.hardware [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.172388] env[62914]: DEBUG nova.virt.hardware [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.172483] env[62914]: DEBUG nova.virt.hardware [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.173354] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c7aa33d-a6c9-40a6-afe1-ba7d55bd06db {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.181934] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e25dfc79-6a87-4ecc-a9fd-d07eee248973 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.196855] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Instance VIF info [] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 604.202751] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Creating folder: Project (399584679d1246d9ab9bc323d441ae62). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 604.203884] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.090s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.208411] env[62914]: ERROR nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 50ee8c92-f158-4391-abba-a9828767fd9b, please check neutron logs for more information. [ 604.208411] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Traceback (most recent call last): [ 604.208411] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.208411] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] self.driver.spawn(context, instance, image_meta, [ 604.208411] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 604.208411] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.208411] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.208411] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] vm_ref = self.build_virtual_machine(instance, [ 604.208411] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.208411] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.208411] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.208687] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] for vif in network_info: [ 604.208687] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.208687] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] return self._sync_wrapper(fn, *args, **kwargs) [ 604.208687] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.208687] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] self.wait() [ 604.208687] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.208687] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] self[:] = self._gt.wait() [ 604.208687] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.208687] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] return self._exit_event.wait() [ 604.208687] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.208687] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] result = hub.switch() [ 604.208687] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.208687] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] return self.greenlet.switch() [ 604.209007] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.209007] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] result = function(*args, **kwargs) [ 604.209007] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.209007] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] return func(*args, **kwargs) [ 604.209007] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.209007] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] raise e [ 604.209007] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.209007] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] nwinfo = self.network_api.allocate_for_instance( [ 604.209007] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.209007] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] created_port_ids = self._update_ports_for_instance( [ 604.209007] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.209007] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] with excutils.save_and_reraise_exception(): [ 604.209007] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.209576] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] self.force_reraise() [ 604.209576] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.209576] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] raise self.value [ 604.209576] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.209576] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] updated_port = self._update_port( [ 604.209576] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.209576] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] _ensure_no_port_binding_failure(port) [ 604.209576] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.209576] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] raise exception.PortBindingFailed(port_id=port['id']) [ 604.209576] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] nova.exception.PortBindingFailed: Binding failed for port 50ee8c92-f158-4391-abba-a9828767fd9b, please check neutron logs for more information. [ 604.209576] env[62914]: ERROR nova.compute.manager [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] [ 604.209914] env[62914]: DEBUG nova.compute.utils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Binding failed for port 50ee8c92-f158-4391-abba-a9828767fd9b, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 604.210459] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7df95d56-7087-41e8-8f23-392f1374732f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.213148] env[62914]: DEBUG nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Build of instance 6ec0c746-2bb8-4387-8b31-b49b0d708614 was re-scheduled: Binding failed for port 50ee8c92-f158-4391-abba-a9828767fd9b, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 604.213148] env[62914]: DEBUG nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 604.213493] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Acquiring lock "refresh_cache-6ec0c746-2bb8-4387-8b31-b49b0d708614" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.213493] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Acquired lock "refresh_cache-6ec0c746-2bb8-4387-8b31-b49b0d708614" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.213661] env[62914]: DEBUG nova.network.neutron [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.215313] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.145s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.216748] env[62914]: INFO nova.compute.claims [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.230021] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Created folder: Project (399584679d1246d9ab9bc323d441ae62) in parent group-v288131. [ 604.230021] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Creating folder: Instances. Parent ref: group-v288148. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 604.230021] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e30232f2-9361-434f-ab7e-68dc2e6ba2cb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.240881] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352372, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063672} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.241424] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 604.243329] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623297a7-16e1-4083-8300-327c844ce5c6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.246023] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Created folder: Instances in parent group-v288148. [ 604.246270] env[62914]: DEBUG oslo.service.loopingcall [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.246470] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 604.246985] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-615dd7cc-6f39-4139-b139-0a087aef7ee0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.276553] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Reconfiguring VM instance instance-00000013 to attach disk [datastore1] f3509973-013e-45ce-87f9-357e782f26d7/f3509973-013e-45ce-87f9-357e782f26d7.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 604.276818] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99081b61-7d0d-4fdc-b382-43587e78ae97 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.292960] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 604.292960] env[62914]: value = "task-1352375" [ 604.292960] env[62914]: _type = "Task" [ 604.292960] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.297601] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 604.297601] env[62914]: value = "task-1352376" [ 604.297601] env[62914]: _type = "Task" [ 604.297601] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.306017] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352375, 'name': CreateVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.309043] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352376, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.812686] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352375, 'name': CreateVM_Task, 'duration_secs': 0.298698} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.815742] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 604.816160] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352376, 'name': ReconfigVM_Task, 'duration_secs': 0.293126} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.816887] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.816887] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.818361] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 604.818361] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Reconfigured VM instance instance-00000013 to attach disk [datastore1] f3509973-013e-45ce-87f9-357e782f26d7/f3509973-013e-45ce-87f9-357e782f26d7.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 604.818361] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33a7fcbb-6313-42a2-a645-384f89ca8341 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.819553] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ab25b75d-65cb-47c4-86b5-61569eb264c0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.824201] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Waiting for the task: (returnval){ [ 604.824201] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]529deaa4-71b1-ded8-c7c5-8b1f1b513d8e" [ 604.824201] env[62914]: _type = "Task" [ 604.824201] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.829821] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 604.829821] env[62914]: value = "task-1352377" [ 604.829821] env[62914]: _type = "Task" [ 604.829821] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.834956] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]529deaa4-71b1-ded8-c7c5-8b1f1b513d8e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.841385] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352377, 'name': Rename_Task} progress is 10%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.923675] env[62914]: DEBUG nova.network.neutron [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.021788] env[62914]: DEBUG nova.network.neutron [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.342937] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352377, 'name': Rename_Task, 'duration_secs': 0.247352} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.349294] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 605.349664] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]529deaa4-71b1-ded8-c7c5-8b1f1b513d8e, 'name': SearchDatastore_Task, 'duration_secs': 0.035936} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.350138] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf7cde26-0db4-441a-a9b4-cf7c92981d15 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.351905] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.352157] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 605.352382] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.352525] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.352696] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 605.353014] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc56be93-7b62-4dcb-ae7f-d32cd75d9ce1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.363703] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 605.363703] env[62914]: value = "task-1352378" [ 605.363703] env[62914]: _type = "Task" [ 605.363703] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.363955] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 605.364131] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 605.365150] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71aa701f-ddf1-4375-b0ee-07df12c97fc5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.375603] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Waiting for the task: (returnval){ [ 605.375603] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f48505-9d46-781b-3207-bd24c2f2dae8" [ 605.375603] env[62914]: _type = "Task" [ 605.375603] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.379292] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352378, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.398736] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f48505-9d46-781b-3207-bd24c2f2dae8, 'name': SearchDatastore_Task, 'duration_secs': 0.015351} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.398736] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffb7bbe0-2030-4012-b8b4-99f2af7e8e2b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.400214] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Waiting for the task: (returnval){ [ 605.400214] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]520a7730-888c-2d97-098f-8815b04c0480" [ 605.400214] env[62914]: _type = "Task" [ 605.400214] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.411293] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]520a7730-888c-2d97-098f-8815b04c0480, 'name': SearchDatastore_Task, 'duration_secs': 0.007937} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.411509] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.411756] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] e11079b9-5f15-41d6-8c8a-c08dc7f63b2f/e11079b9-5f15-41d6-8c8a-c08dc7f63b2f.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 605.412009] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-191b0e23-10f9-4229-abac-198a125cf29a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.418224] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Waiting for the task: (returnval){ [ 605.418224] env[62914]: value = "task-1352379" [ 605.418224] env[62914]: _type = "Task" [ 605.418224] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.426059] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352379, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.523132] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Releasing lock "refresh_cache-6ec0c746-2bb8-4387-8b31-b49b0d708614" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.523374] env[62914]: DEBUG nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 605.523956] env[62914]: DEBUG nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.523956] env[62914]: DEBUG nova.network.neutron [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 605.546271] env[62914]: DEBUG nova.network.neutron [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.814455] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85244453-9255-4889-8dbb-49ee63cb0ac2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.824894] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a70c8c-4387-4faa-9110-9c3bf10f29e3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.863354] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe791429-b39f-419b-9d58-9e14d746b8a0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.878988] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc992b1-6b76-453b-9bee-17efb90da11a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.881347] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352378, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.892366] env[62914]: DEBUG nova.compute.provider_tree [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.929805] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352379, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.049709] env[62914]: DEBUG nova.network.neutron [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.375663] env[62914]: DEBUG oslo_vmware.api [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352378, 'name': PowerOnVM_Task, 'duration_secs': 0.615892} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.375962] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 606.376449] env[62914]: INFO nova.compute.manager [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Took 4.88 seconds to spawn the instance on the hypervisor. [ 606.376657] env[62914]: DEBUG nova.compute.manager [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 606.377456] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2020d44-9427-4ea3-9ec6-b9846d46561f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.398989] env[62914]: DEBUG nova.scheduler.client.report [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.429139] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352379, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.586884} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.429453] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] e11079b9-5f15-41d6-8c8a-c08dc7f63b2f/e11079b9-5f15-41d6-8c8a-c08dc7f63b2f.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 606.429802] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 606.430095] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f4495b28-4765-4b92-b334-c38ec29cf572 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.437086] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Waiting for the task: (returnval){ [ 606.437086] env[62914]: value = "task-1352380" [ 606.437086] env[62914]: _type = "Task" [ 606.437086] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.445458] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352380, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.554909] env[62914]: INFO nova.compute.manager [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] [instance: 6ec0c746-2bb8-4387-8b31-b49b0d708614] Took 1.03 seconds to deallocate network for instance. [ 606.896597] env[62914]: INFO nova.compute.manager [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Took 31.61 seconds to build instance. [ 606.900221] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.685s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.900707] env[62914]: DEBUG nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 606.903983] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.036s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.905253] env[62914]: INFO nova.compute.claims [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.948935] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352380, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062203} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.949257] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 606.950281] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7780599-f7e2-4caf-9259-e7ada19989a4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.978047] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Reconfiguring VM instance instance-00000014 to attach disk [datastore2] e11079b9-5f15-41d6-8c8a-c08dc7f63b2f/e11079b9-5f15-41d6-8c8a-c08dc7f63b2f.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 606.980034] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-082aff1a-e4bd-4c99-ae78-aecd1f3e8873 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.001866] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Waiting for the task: (returnval){ [ 607.001866] env[62914]: value = "task-1352381" [ 607.001866] env[62914]: _type = "Task" [ 607.001866] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.009995] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352381, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.232425] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "98c19874-7c71-4e20-aa29-2ddc53df7e81" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.232628] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "98c19874-7c71-4e20-aa29-2ddc53df7e81" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.277106] env[62914]: INFO nova.compute.manager [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Rebuilding instance [ 607.326622] env[62914]: DEBUG nova.compute.manager [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 607.327496] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea00510-227b-4b67-9e86-df449fcb5130 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.398856] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2859a9ea-6729-4457-8c2a-3bb5d3b37184 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "f3509973-013e-45ce-87f9-357e782f26d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.405s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.409695] env[62914]: DEBUG nova.compute.utils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.415022] env[62914]: DEBUG nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.415022] env[62914]: DEBUG nova.network.neutron [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 607.491373] env[62914]: DEBUG nova.policy [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '961f770e6dd94b74ac253cb8fd6e43cd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c09e038804814fe4aef75c79fa652413', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 607.513457] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352381, 'name': ReconfigVM_Task, 'duration_secs': 0.264352} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.513736] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Reconfigured VM instance instance-00000014 to attach disk [datastore2] e11079b9-5f15-41d6-8c8a-c08dc7f63b2f/e11079b9-5f15-41d6-8c8a-c08dc7f63b2f.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 607.514393] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dfbe2bc6-4487-46bb-858e-d576e7792b8e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.521950] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Waiting for the task: (returnval){ [ 607.521950] env[62914]: value = "task-1352382" [ 607.521950] env[62914]: _type = "Task" [ 607.521950] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.529733] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352382, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.597628] env[62914]: INFO nova.scheduler.client.report [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Deleted allocations for instance 6ec0c746-2bb8-4387-8b31-b49b0d708614 [ 607.839413] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 607.839726] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f644b8f9-7102-41a7-9681-98124bc887f2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.846699] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 607.846699] env[62914]: value = "task-1352383" [ 607.846699] env[62914]: _type = "Task" [ 607.846699] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.855407] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352383, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.895389] env[62914]: DEBUG nova.network.neutron [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Successfully created port: 1194b118-455c-4a76-a106-cde6931bec18 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 607.902644] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 607.917742] env[62914]: DEBUG nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 608.031494] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352382, 'name': Rename_Task, 'duration_secs': 0.136724} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.038204] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 608.038653] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf47abfd-3574-4ca8-85dd-775470b42571 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.045154] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Waiting for the task: (returnval){ [ 608.045154] env[62914]: value = "task-1352384" [ 608.045154] env[62914]: _type = "Task" [ 608.045154] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.053256] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352384, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.107096] env[62914]: DEBUG oslo_concurrency.lockutils [None req-18b3f57a-b6f3-455a-b37f-6bfd0c472f28 tempest-FloatingIPsAssociationNegativeTestJSON-797914257 tempest-FloatingIPsAssociationNegativeTestJSON-797914257-project-member] Lock "6ec0c746-2bb8-4387-8b31-b49b0d708614" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.888s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.360464] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352383, 'name': PowerOffVM_Task, 'duration_secs': 0.140199} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.360915] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 608.360965] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 608.362356] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de215169-4633-4446-8c38-6cd7c2ddf76e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.369587] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 608.369893] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-50c92ef9-49bf-4074-981a-cb187508c8a5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.395663] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 608.395875] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 608.396071] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Deleting the datastore file [datastore1] f3509973-013e-45ce-87f9-357e782f26d7 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 608.396347] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f8d438b-eee6-4a6e-961a-95d8740e9f53 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.417248] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 608.417248] env[62914]: value = "task-1352386" [ 608.417248] env[62914]: _type = "Task" [ 608.417248] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.422735] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352386, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.432362] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.509870] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae19a130-ae24-4646-b756-691934b5711a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.518874] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081c2257-5fa5-40f3-8a48-1f1cd197a150 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.556571] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d157f4-a0c3-4c5d-989f-916a50c1204e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.564991] env[62914]: DEBUG oslo_vmware.api [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352384, 'name': PowerOnVM_Task, 'duration_secs': 0.438822} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.567055] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 608.567347] env[62914]: INFO nova.compute.manager [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Took 4.43 seconds to spawn the instance on the hypervisor. [ 608.567443] env[62914]: DEBUG nova.compute.manager [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 608.568273] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2681c872-b9a9-4df2-8fc2-d9448c7630f8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.575021] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fee16b-a719-439c-b149-2d7a58143671 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.589084] env[62914]: DEBUG nova.compute.provider_tree [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.611657] env[62914]: DEBUG nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 608.818282] env[62914]: DEBUG nova.compute.manager [req-351bb399-35de-4655-8e44-47ad9f1cb0be req-3ff25503-d173-43e1-955a-5e4a12409ac8 service nova] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Received event network-changed-1194b118-455c-4a76-a106-cde6931bec18 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.818517] env[62914]: DEBUG nova.compute.manager [req-351bb399-35de-4655-8e44-47ad9f1cb0be req-3ff25503-d173-43e1-955a-5e4a12409ac8 service nova] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Refreshing instance network info cache due to event network-changed-1194b118-455c-4a76-a106-cde6931bec18. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 608.818730] env[62914]: DEBUG oslo_concurrency.lockutils [req-351bb399-35de-4655-8e44-47ad9f1cb0be req-3ff25503-d173-43e1-955a-5e4a12409ac8 service nova] Acquiring lock "refresh_cache-8213bc31-1bd5-40b6-99a0-1254b9517b6a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.818868] env[62914]: DEBUG oslo_concurrency.lockutils [req-351bb399-35de-4655-8e44-47ad9f1cb0be req-3ff25503-d173-43e1-955a-5e4a12409ac8 service nova] Acquired lock "refresh_cache-8213bc31-1bd5-40b6-99a0-1254b9517b6a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.819862] env[62914]: DEBUG nova.network.neutron [req-351bb399-35de-4655-8e44-47ad9f1cb0be req-3ff25503-d173-43e1-955a-5e4a12409ac8 service nova] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Refreshing network info cache for port 1194b118-455c-4a76-a106-cde6931bec18 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 608.924839] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352386, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130348} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.925198] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 608.925419] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 608.925597] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 608.937817] env[62914]: DEBUG nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 608.974221] env[62914]: DEBUG nova.virt.hardware [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 608.974480] env[62914]: DEBUG nova.virt.hardware [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 608.974634] env[62914]: DEBUG nova.virt.hardware [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 608.974816] env[62914]: DEBUG nova.virt.hardware [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 608.974957] env[62914]: DEBUG nova.virt.hardware [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 608.975118] env[62914]: DEBUG nova.virt.hardware [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 608.975371] env[62914]: DEBUG nova.virt.hardware [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 608.975547] env[62914]: DEBUG nova.virt.hardware [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 608.975714] env[62914]: DEBUG nova.virt.hardware [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 608.975873] env[62914]: DEBUG nova.virt.hardware [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 608.976051] env[62914]: DEBUG nova.virt.hardware [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.976914] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41332036-c89a-4778-b920-b44116c68723 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.986192] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467e198d-49bf-4cfa-b723-c643a4993175 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.037885] env[62914]: ERROR nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1194b118-455c-4a76-a106-cde6931bec18, please check neutron logs for more information. [ 609.037885] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 609.037885] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.037885] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 609.037885] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.037885] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 609.037885] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.037885] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 609.037885] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.037885] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 609.037885] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.037885] env[62914]: ERROR nova.compute.manager raise self.value [ 609.037885] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.037885] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 609.037885] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.037885] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 609.038355] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.038355] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 609.038355] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1194b118-455c-4a76-a106-cde6931bec18, please check neutron logs for more information. [ 609.038355] env[62914]: ERROR nova.compute.manager [ 609.038355] env[62914]: Traceback (most recent call last): [ 609.038355] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 609.038355] env[62914]: listener.cb(fileno) [ 609.038355] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.038355] env[62914]: result = function(*args, **kwargs) [ 609.038355] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 609.038355] env[62914]: return func(*args, **kwargs) [ 609.038355] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.038355] env[62914]: raise e [ 609.038355] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.038355] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 609.038355] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.038355] env[62914]: created_port_ids = self._update_ports_for_instance( [ 609.038355] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.038355] env[62914]: with excutils.save_and_reraise_exception(): [ 609.038355] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.038355] env[62914]: self.force_reraise() [ 609.038355] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.038355] env[62914]: raise self.value [ 609.038355] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.038355] env[62914]: updated_port = self._update_port( [ 609.038355] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.038355] env[62914]: _ensure_no_port_binding_failure(port) [ 609.038355] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.038355] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 609.039083] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 1194b118-455c-4a76-a106-cde6931bec18, please check neutron logs for more information. [ 609.039083] env[62914]: Removing descriptor: 19 [ 609.039083] env[62914]: ERROR nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1194b118-455c-4a76-a106-cde6931bec18, please check neutron logs for more information. [ 609.039083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Traceback (most recent call last): [ 609.039083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 609.039083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] yield resources [ 609.039083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.039083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] self.driver.spawn(context, instance, image_meta, [ 609.039083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 609.039083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.039083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.039083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] vm_ref = self.build_virtual_machine(instance, [ 609.039401] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.039401] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.039401] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.039401] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] for vif in network_info: [ 609.039401] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.039401] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] return self._sync_wrapper(fn, *args, **kwargs) [ 609.039401] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.039401] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] self.wait() [ 609.039401] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.039401] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] self[:] = self._gt.wait() [ 609.039401] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.039401] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] return self._exit_event.wait() [ 609.039401] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.039736] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] result = hub.switch() [ 609.039736] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.039736] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] return self.greenlet.switch() [ 609.039736] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.039736] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] result = function(*args, **kwargs) [ 609.039736] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 609.039736] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] return func(*args, **kwargs) [ 609.039736] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.039736] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] raise e [ 609.039736] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.039736] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] nwinfo = self.network_api.allocate_for_instance( [ 609.039736] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.039736] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] created_port_ids = self._update_ports_for_instance( [ 609.040083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.040083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] with excutils.save_and_reraise_exception(): [ 609.040083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.040083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] self.force_reraise() [ 609.040083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.040083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] raise self.value [ 609.040083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.040083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] updated_port = self._update_port( [ 609.040083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.040083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] _ensure_no_port_binding_failure(port) [ 609.040083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.040083] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] raise exception.PortBindingFailed(port_id=port['id']) [ 609.040412] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] nova.exception.PortBindingFailed: Binding failed for port 1194b118-455c-4a76-a106-cde6931bec18, please check neutron logs for more information. [ 609.040412] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] [ 609.040412] env[62914]: INFO nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Terminating instance [ 609.043017] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Acquiring lock "refresh_cache-8213bc31-1bd5-40b6-99a0-1254b9517b6a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.103022] env[62914]: DEBUG nova.scheduler.client.report [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.103022] env[62914]: INFO nova.compute.manager [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Took 29.32 seconds to build instance. [ 609.142037] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.801034] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.897s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.801487] env[62914]: DEBUG nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 609.803852] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b6ee14-840e-4d5c-a956-2196cd09f6ea tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Lock "e11079b9-5f15-41d6-8c8a-c08dc7f63b2f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.209s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.804421] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.442s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.822300] env[62914]: DEBUG nova.network.neutron [req-351bb399-35de-4655-8e44-47ad9f1cb0be req-3ff25503-d173-43e1-955a-5e4a12409ac8 service nova] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.892322] env[62914]: DEBUG nova.network.neutron [req-351bb399-35de-4655-8e44-47ad9f1cb0be req-3ff25503-d173-43e1-955a-5e4a12409ac8 service nova] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.306364] env[62914]: DEBUG nova.compute.utils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 610.310388] env[62914]: DEBUG nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 610.310559] env[62914]: DEBUG nova.network.neutron [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 610.314787] env[62914]: DEBUG nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 610.338202] env[62914]: DEBUG nova.virt.hardware [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.338202] env[62914]: DEBUG nova.virt.hardware [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.338202] env[62914]: DEBUG nova.virt.hardware [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.338202] env[62914]: DEBUG nova.virt.hardware [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.338458] env[62914]: DEBUG nova.virt.hardware [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.338458] env[62914]: DEBUG nova.virt.hardware [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.338611] env[62914]: DEBUG nova.virt.hardware [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.338771] env[62914]: DEBUG nova.virt.hardware [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.338898] env[62914]: DEBUG nova.virt.hardware [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.339071] env[62914]: DEBUG nova.virt.hardware [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.339242] env[62914]: DEBUG nova.virt.hardware [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.340861] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6414bda-cedf-40b4-8521-c5368db6ffa5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.351190] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1498e26f-559c-43ad-aa00-4254b15e03a6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.365464] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Instance VIF info [] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 610.371033] env[62914]: DEBUG oslo.service.loopingcall [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 610.371876] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 610.372113] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c38e816a-ab20-4136-a56e-bd34b566bb2d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.387784] env[62914]: DEBUG nova.policy [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '322926325d12480c947dba38fda00899', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f48fa131f322400cb28c969ceb9b7cc2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 610.394681] env[62914]: DEBUG oslo_concurrency.lockutils [req-351bb399-35de-4655-8e44-47ad9f1cb0be req-3ff25503-d173-43e1-955a-5e4a12409ac8 service nova] Releasing lock "refresh_cache-8213bc31-1bd5-40b6-99a0-1254b9517b6a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.395914] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Acquired lock "refresh_cache-8213bc31-1bd5-40b6-99a0-1254b9517b6a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.396120] env[62914]: DEBUG nova.network.neutron [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.397320] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 610.397320] env[62914]: value = "task-1352387" [ 610.397320] env[62914]: _type = "Task" [ 610.397320] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.407638] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352387, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.627974] env[62914]: DEBUG nova.compute.manager [None req-6209dec8-35c9-4533-8f81-235babef2acd tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 610.628823] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17b878b-2419-4e54-9d51-0367177ad225 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.748096] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Acquiring lock "e11079b9-5f15-41d6-8c8a-c08dc7f63b2f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.748459] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Lock "e11079b9-5f15-41d6-8c8a-c08dc7f63b2f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.750989] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Acquiring lock "e11079b9-5f15-41d6-8c8a-c08dc7f63b2f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.750989] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Lock "e11079b9-5f15-41d6-8c8a-c08dc7f63b2f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.750989] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Lock "e11079b9-5f15-41d6-8c8a-c08dc7f63b2f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.755355] env[62914]: INFO nova.compute.manager [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Terminating instance [ 610.755355] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Acquiring lock "refresh_cache-e11079b9-5f15-41d6-8c8a-c08dc7f63b2f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.755355] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Acquired lock "refresh_cache-e11079b9-5f15-41d6-8c8a-c08dc7f63b2f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.755355] env[62914]: DEBUG nova.network.neutron [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.801647] env[62914]: DEBUG nova.network.neutron [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Successfully created port: 691b9a22-1a0a-4a84-a136-b828743aff9c {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 610.812044] env[62914]: DEBUG nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 610.843577] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.876775] env[62914]: DEBUG nova.compute.manager [req-b904b88e-2f28-4561-a8d5-1a5ea85a84b4 req-71f1f1b0-70ba-41f8-b122-590bd6de56ec service nova] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Received event network-vif-deleted-1194b118-455c-4a76-a106-cde6931bec18 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 610.908993] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c8af1e-5faf-4064-bb5c-658ff3fac6a6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.915693] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352387, 'name': CreateVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.920098] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca182d4d-8ab8-4378-9d7f-9bc5be2bb7de {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.955736] env[62914]: DEBUG nova.network.neutron [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.958586] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8f9fc3-a144-4fab-9d72-492fa2f9c599 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.965414] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311ae783-5a79-410f-a568-9f5a04f2a7fa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.980768] env[62914]: DEBUG nova.compute.provider_tree [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.031208] env[62914]: DEBUG nova.network.neutron [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.145114] env[62914]: INFO nova.compute.manager [None req-6209dec8-35c9-4533-8f81-235babef2acd tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] instance snapshotting [ 611.145824] env[62914]: DEBUG nova.objects.instance [None req-6209dec8-35c9-4533-8f81-235babef2acd tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Lazy-loading 'flavor' on Instance uuid e11079b9-5f15-41d6-8c8a-c08dc7f63b2f {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 611.284463] env[62914]: DEBUG nova.network.neutron [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.344020] env[62914]: DEBUG nova.network.neutron [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.410024] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352387, 'name': CreateVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.488511] env[62914]: DEBUG nova.scheduler.client.report [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.534129] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Releasing lock "refresh_cache-8213bc31-1bd5-40b6-99a0-1254b9517b6a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.534544] env[62914]: DEBUG nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 611.534747] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 611.535211] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b7ba9d7-9101-4ed6-9c17-18c120962bf5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.544341] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3dbf810-bffa-4b82-8cad-6a7c2e9ba4e8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.570106] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8213bc31-1bd5-40b6-99a0-1254b9517b6a could not be found. [ 611.570403] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 611.570622] env[62914]: INFO nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 611.570884] env[62914]: DEBUG oslo.service.loopingcall [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.571150] env[62914]: DEBUG nova.compute.manager [-] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.571289] env[62914]: DEBUG nova.network.neutron [-] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 611.600621] env[62914]: DEBUG nova.network.neutron [-] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.652479] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1720a7a1-54dc-4f2d-b154-7fd3dcf2c127 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.670437] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de626989-3aeb-4cf4-9f62-fe5b36728f11 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.825516] env[62914]: DEBUG nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 611.844525] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Releasing lock "refresh_cache-e11079b9-5f15-41d6-8c8a-c08dc7f63b2f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.845097] env[62914]: DEBUG nova.compute.manager [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 611.845519] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 611.846827] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86038823-6be9-47bb-abdb-ea6e8446305a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.852410] env[62914]: DEBUG nova.virt.hardware [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 611.852670] env[62914]: DEBUG nova.virt.hardware [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 611.852827] env[62914]: DEBUG nova.virt.hardware [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 611.853011] env[62914]: DEBUG nova.virt.hardware [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 611.853181] env[62914]: DEBUG nova.virt.hardware [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 611.853365] env[62914]: DEBUG nova.virt.hardware [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 611.853575] env[62914]: DEBUG nova.virt.hardware [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 611.853724] env[62914]: DEBUG nova.virt.hardware [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 611.853880] env[62914]: DEBUG nova.virt.hardware [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 611.854049] env[62914]: DEBUG nova.virt.hardware [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 611.854217] env[62914]: DEBUG nova.virt.hardware [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 611.855131] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db49a4ad-c909-44db-b507-61337ab0235b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.859931] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 611.860501] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c373351-9413-492c-8701-8fc0913316f2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.865063] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07af0521-a6b1-4423-946e-c50e9b83c10c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.870664] env[62914]: DEBUG oslo_vmware.api [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Waiting for the task: (returnval){ [ 611.870664] env[62914]: value = "task-1352388" [ 611.870664] env[62914]: _type = "Task" [ 611.870664] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.887556] env[62914]: DEBUG oslo_vmware.api [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352388, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.900241] env[62914]: ERROR nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 691b9a22-1a0a-4a84-a136-b828743aff9c, please check neutron logs for more information. [ 611.900241] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 611.900241] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.900241] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 611.900241] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.900241] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 611.900241] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.900241] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 611.900241] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.900241] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 611.900241] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.900241] env[62914]: ERROR nova.compute.manager raise self.value [ 611.900241] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.900241] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 611.900241] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.900241] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 611.900628] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.900628] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 611.900628] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 691b9a22-1a0a-4a84-a136-b828743aff9c, please check neutron logs for more information. [ 611.900628] env[62914]: ERROR nova.compute.manager [ 611.900628] env[62914]: Traceback (most recent call last): [ 611.900628] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 611.900628] env[62914]: listener.cb(fileno) [ 611.900628] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.900628] env[62914]: result = function(*args, **kwargs) [ 611.900628] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.900628] env[62914]: return func(*args, **kwargs) [ 611.900628] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.900628] env[62914]: raise e [ 611.900628] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.900628] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 611.900628] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.900628] env[62914]: created_port_ids = self._update_ports_for_instance( [ 611.900628] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.900628] env[62914]: with excutils.save_and_reraise_exception(): [ 611.900628] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.900628] env[62914]: self.force_reraise() [ 611.900628] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.900628] env[62914]: raise self.value [ 611.900628] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.900628] env[62914]: updated_port = self._update_port( [ 611.900628] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.900628] env[62914]: _ensure_no_port_binding_failure(port) [ 611.900628] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.900628] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 611.901862] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 691b9a22-1a0a-4a84-a136-b828743aff9c, please check neutron logs for more information. [ 611.901862] env[62914]: Removing descriptor: 19 [ 611.901862] env[62914]: ERROR nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 691b9a22-1a0a-4a84-a136-b828743aff9c, please check neutron logs for more information. [ 611.901862] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Traceback (most recent call last): [ 611.901862] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 611.901862] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] yield resources [ 611.901862] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.901862] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] self.driver.spawn(context, instance, image_meta, [ 611.901862] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 611.901862] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.901862] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.901862] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] vm_ref = self.build_virtual_machine(instance, [ 611.902153] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.902153] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.902153] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.902153] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] for vif in network_info: [ 611.902153] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.902153] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] return self._sync_wrapper(fn, *args, **kwargs) [ 611.902153] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.902153] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] self.wait() [ 611.902153] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.902153] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] self[:] = self._gt.wait() [ 611.902153] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.902153] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] return self._exit_event.wait() [ 611.902153] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.902431] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] result = hub.switch() [ 611.902431] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.902431] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] return self.greenlet.switch() [ 611.902431] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.902431] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] result = function(*args, **kwargs) [ 611.902431] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.902431] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] return func(*args, **kwargs) [ 611.902431] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.902431] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] raise e [ 611.902431] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.902431] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] nwinfo = self.network_api.allocate_for_instance( [ 611.902431] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.902431] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] created_port_ids = self._update_ports_for_instance( [ 611.902716] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.902716] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] with excutils.save_and_reraise_exception(): [ 611.902716] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.902716] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] self.force_reraise() [ 611.902716] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.902716] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] raise self.value [ 611.902716] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.902716] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] updated_port = self._update_port( [ 611.902716] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.902716] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] _ensure_no_port_binding_failure(port) [ 611.902716] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.902716] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] raise exception.PortBindingFailed(port_id=port['id']) [ 611.902990] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] nova.exception.PortBindingFailed: Binding failed for port 691b9a22-1a0a-4a84-a136-b828743aff9c, please check neutron logs for more information. [ 611.902990] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] [ 611.902990] env[62914]: INFO nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Terminating instance [ 611.906714] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Acquiring lock "refresh_cache-5e695914-afa5-4f79-972d-852a4ac8aa20" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.906879] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Acquired lock "refresh_cache-5e695914-afa5-4f79-972d-852a4ac8aa20" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.907053] env[62914]: DEBUG nova.network.neutron [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.914848] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352387, 'name': CreateVM_Task, 'duration_secs': 1.335345} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.914848] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 611.914848] env[62914]: DEBUG oslo_concurrency.lockutils [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.915062] env[62914]: DEBUG oslo_concurrency.lockutils [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.915495] env[62914]: DEBUG oslo_concurrency.lockutils [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 611.916111] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc594f37-8ff4-4bb2-abe7-e256102f85c9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.921532] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 611.921532] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52498fa7-a242-30f8-4ea1-f5bfd367cba9" [ 611.921532] env[62914]: _type = "Task" [ 611.921532] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.930665] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52498fa7-a242-30f8-4ea1-f5bfd367cba9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.994096] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.189s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.994833] env[62914]: ERROR nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ef6080b9-52f1-4d13-8cf4-b79ec17bbdf2, please check neutron logs for more information. [ 611.994833] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] Traceback (most recent call last): [ 611.994833] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.994833] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] self.driver.spawn(context, instance, image_meta, [ 611.994833] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 611.994833] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.994833] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.994833] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] vm_ref = self.build_virtual_machine(instance, [ 611.994833] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.994833] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.994833] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.995112] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] for vif in network_info: [ 611.995112] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.995112] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] return self._sync_wrapper(fn, *args, **kwargs) [ 611.995112] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.995112] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] self.wait() [ 611.995112] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.995112] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] self[:] = self._gt.wait() [ 611.995112] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.995112] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] return self._exit_event.wait() [ 611.995112] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.995112] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] result = hub.switch() [ 611.995112] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.995112] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] return self.greenlet.switch() [ 611.995437] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.995437] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] result = function(*args, **kwargs) [ 611.995437] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.995437] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] return func(*args, **kwargs) [ 611.995437] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.995437] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] raise e [ 611.995437] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.995437] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] nwinfo = self.network_api.allocate_for_instance( [ 611.995437] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.995437] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] created_port_ids = self._update_ports_for_instance( [ 611.995437] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.995437] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] with excutils.save_and_reraise_exception(): [ 611.995437] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.995741] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] self.force_reraise() [ 611.995741] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.995741] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] raise self.value [ 611.995741] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.995741] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] updated_port = self._update_port( [ 611.995741] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.995741] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] _ensure_no_port_binding_failure(port) [ 611.995741] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.995741] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] raise exception.PortBindingFailed(port_id=port['id']) [ 611.995741] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] nova.exception.PortBindingFailed: Binding failed for port ef6080b9-52f1-4d13-8cf4-b79ec17bbdf2, please check neutron logs for more information. [ 611.995741] env[62914]: ERROR nova.compute.manager [instance: 723105bc-6810-4759-8240-af0a128475e4] [ 611.995985] env[62914]: DEBUG nova.compute.utils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Binding failed for port ef6080b9-52f1-4d13-8cf4-b79ec17bbdf2, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 611.997161] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.411s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.998633] env[62914]: INFO nova.compute.claims [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.001368] env[62914]: DEBUG nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Build of instance 723105bc-6810-4759-8240-af0a128475e4 was re-scheduled: Binding failed for port ef6080b9-52f1-4d13-8cf4-b79ec17bbdf2, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 612.001850] env[62914]: DEBUG nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 612.002089] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Acquiring lock "refresh_cache-723105bc-6810-4759-8240-af0a128475e4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.002301] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Acquired lock "refresh_cache-723105bc-6810-4759-8240-af0a128475e4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.002485] env[62914]: DEBUG nova.network.neutron [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.103608] env[62914]: DEBUG nova.network.neutron [-] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.180857] env[62914]: DEBUG nova.compute.manager [None req-6209dec8-35c9-4533-8f81-235babef2acd tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Instance disappeared during snapshot {{(pid=62914) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 612.312585] env[62914]: DEBUG nova.compute.manager [None req-6209dec8-35c9-4533-8f81-235babef2acd tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Found 0 images (rotation: 2) {{(pid=62914) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 612.381301] env[62914]: DEBUG oslo_vmware.api [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352388, 'name': PowerOffVM_Task, 'duration_secs': 0.199642} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.381586] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 612.381751] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 612.382007] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9fb2cfeb-ee55-47ff-a860-2c353ca7de04 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.406308] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 612.406308] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 612.406308] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Deleting the datastore file [datastore2] e11079b9-5f15-41d6-8c8a-c08dc7f63b2f {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 612.406617] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1120f6a0-cb9f-48da-9f53-3dca36236103 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.413360] env[62914]: DEBUG oslo_vmware.api [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Waiting for the task: (returnval){ [ 612.413360] env[62914]: value = "task-1352390" [ 612.413360] env[62914]: _type = "Task" [ 612.413360] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.422321] env[62914]: DEBUG oslo_vmware.api [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352390, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.431172] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52498fa7-a242-30f8-4ea1-f5bfd367cba9, 'name': SearchDatastore_Task, 'duration_secs': 0.011239} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.432106] env[62914]: DEBUG nova.network.neutron [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.433925] env[62914]: DEBUG oslo_concurrency.lockutils [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.434186] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 612.434447] env[62914]: DEBUG oslo_concurrency.lockutils [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.434582] env[62914]: DEBUG oslo_concurrency.lockutils [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.434785] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 612.435075] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9c56e21b-827b-477d-bb27-683855983d4f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.443990] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 612.444204] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 612.445061] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd23f9b4-b967-4ada-9ee1-97d2e9841fe7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.450199] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 612.450199] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52dd6921-fa29-13e1-d1b4-f805abddf1c5" [ 612.450199] env[62914]: _type = "Task" [ 612.450199] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.459729] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52dd6921-fa29-13e1-d1b4-f805abddf1c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.531271] env[62914]: DEBUG nova.network.neutron [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.592337] env[62914]: DEBUG nova.network.neutron [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.605794] env[62914]: INFO nova.compute.manager [-] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Took 1.03 seconds to deallocate network for instance. [ 612.609045] env[62914]: DEBUG nova.compute.claims [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 612.609234] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.622148] env[62914]: DEBUG nova.network.neutron [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.900281] env[62914]: DEBUG nova.compute.manager [req-b21dedc2-0642-415f-b75f-5177d0b2e52c req-14cab4b9-22ec-478d-ad90-4aa3251849d1 service nova] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Received event network-changed-691b9a22-1a0a-4a84-a136-b828743aff9c {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.900442] env[62914]: DEBUG nova.compute.manager [req-b21dedc2-0642-415f-b75f-5177d0b2e52c req-14cab4b9-22ec-478d-ad90-4aa3251849d1 service nova] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Refreshing instance network info cache due to event network-changed-691b9a22-1a0a-4a84-a136-b828743aff9c. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 612.900631] env[62914]: DEBUG oslo_concurrency.lockutils [req-b21dedc2-0642-415f-b75f-5177d0b2e52c req-14cab4b9-22ec-478d-ad90-4aa3251849d1 service nova] Acquiring lock "refresh_cache-5e695914-afa5-4f79-972d-852a4ac8aa20" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.922811] env[62914]: DEBUG oslo_vmware.api [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Task: {'id': task-1352390, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099241} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.923055] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 612.923255] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 612.923433] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 612.923610] env[62914]: INFO nova.compute.manager [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Took 1.08 seconds to destroy the instance on the hypervisor. [ 612.923839] env[62914]: DEBUG oslo.service.loopingcall [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.924073] env[62914]: DEBUG nova.compute.manager [-] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.924197] env[62914]: DEBUG nova.network.neutron [-] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.938925] env[62914]: DEBUG nova.network.neutron [-] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.959976] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52dd6921-fa29-13e1-d1b4-f805abddf1c5, 'name': SearchDatastore_Task, 'duration_secs': 0.00876} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.960745] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ee9a119-c4b6-4e66-8168-fd3e61b543d8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.966232] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 612.966232] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5228796e-53f6-0415-8f02-769ccc4b2609" [ 612.966232] env[62914]: _type = "Task" [ 612.966232] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.974175] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5228796e-53f6-0415-8f02-769ccc4b2609, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.095776] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Releasing lock "refresh_cache-5e695914-afa5-4f79-972d-852a4ac8aa20" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.096197] env[62914]: DEBUG nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 613.096411] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 613.096896] env[62914]: DEBUG oslo_concurrency.lockutils [req-b21dedc2-0642-415f-b75f-5177d0b2e52c req-14cab4b9-22ec-478d-ad90-4aa3251849d1 service nova] Acquired lock "refresh_cache-5e695914-afa5-4f79-972d-852a4ac8aa20" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.097084] env[62914]: DEBUG nova.network.neutron [req-b21dedc2-0642-415f-b75f-5177d0b2e52c req-14cab4b9-22ec-478d-ad90-4aa3251849d1 service nova] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Refreshing network info cache for port 691b9a22-1a0a-4a84-a136-b828743aff9c {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 613.098379] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-02b98d01-7734-4960-acff-19f10a1af77a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.107165] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d9cffe-3f36-430b-9940-0cf7e740fe65 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.124198] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Releasing lock "refresh_cache-723105bc-6810-4759-8240-af0a128475e4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.124518] env[62914]: DEBUG nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 613.124805] env[62914]: DEBUG nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.125098] env[62914]: DEBUG nova.network.neutron [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.134029] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5e695914-afa5-4f79-972d-852a4ac8aa20 could not be found. [ 613.134247] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 613.134422] env[62914]: INFO nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Took 0.04 seconds to destroy the instance on the hypervisor. [ 613.134658] env[62914]: DEBUG oslo.service.loopingcall [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.137058] env[62914]: DEBUG nova.compute.manager [-] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.137152] env[62914]: DEBUG nova.network.neutron [-] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.146720] env[62914]: DEBUG nova.network.neutron [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.162503] env[62914]: DEBUG nova.network.neutron [-] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.441402] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b7ab47-ad2d-4021-b71e-685ffb7c30f0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.444008] env[62914]: DEBUG nova.network.neutron [-] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.451093] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25747b57-705e-4f91-a5ed-8bcfb7535932 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.486233] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63152882-2fc3-4467-afc2-ce73ddb609e6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.494598] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5228796e-53f6-0415-8f02-769ccc4b2609, 'name': SearchDatastore_Task, 'duration_secs': 0.009616} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.496653] env[62914]: DEBUG oslo_concurrency.lockutils [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.496913] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] f3509973-013e-45ce-87f9-357e782f26d7/f3509973-013e-45ce-87f9-357e782f26d7.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 613.497477] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5019d4e8-2162-4b9a-9bd7-2e58e6b256a2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.500342] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c5f7a7-5a16-4d5c-b2b6-3b58ccaa74b6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.516929] env[62914]: DEBUG nova.compute.provider_tree [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.519569] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 613.519569] env[62914]: value = "task-1352391" [ 613.519569] env[62914]: _type = "Task" [ 613.519569] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.527694] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352391, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.626836] env[62914]: DEBUG nova.network.neutron [req-b21dedc2-0642-415f-b75f-5177d0b2e52c req-14cab4b9-22ec-478d-ad90-4aa3251849d1 service nova] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.652744] env[62914]: DEBUG nova.network.neutron [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.665161] env[62914]: DEBUG nova.network.neutron [-] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.934935] env[62914]: DEBUG nova.network.neutron [req-b21dedc2-0642-415f-b75f-5177d0b2e52c req-14cab4b9-22ec-478d-ad90-4aa3251849d1 service nova] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.950034] env[62914]: INFO nova.compute.manager [-] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Took 1.02 seconds to deallocate network for instance. [ 614.022280] env[62914]: DEBUG nova.scheduler.client.report [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.035677] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352391, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485391} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.035966] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] f3509973-013e-45ce-87f9-357e782f26d7/f3509973-013e-45ce-87f9-357e782f26d7.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 614.036144] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 614.036385] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea9279bb-4f1d-40bc-a5d7-c03e93a05fbe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.043147] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 614.043147] env[62914]: value = "task-1352392" [ 614.043147] env[62914]: _type = "Task" [ 614.043147] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.051180] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352392, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.154856] env[62914]: INFO nova.compute.manager [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] [instance: 723105bc-6810-4759-8240-af0a128475e4] Took 1.03 seconds to deallocate network for instance. [ 614.166437] env[62914]: INFO nova.compute.manager [-] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Took 1.03 seconds to deallocate network for instance. [ 614.168546] env[62914]: DEBUG nova.compute.claims [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 614.168721] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.437267] env[62914]: DEBUG oslo_concurrency.lockutils [req-b21dedc2-0642-415f-b75f-5177d0b2e52c req-14cab4b9-22ec-478d-ad90-4aa3251849d1 service nova] Releasing lock "refresh_cache-5e695914-afa5-4f79-972d-852a4ac8aa20" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.437525] env[62914]: DEBUG nova.compute.manager [req-b21dedc2-0642-415f-b75f-5177d0b2e52c req-14cab4b9-22ec-478d-ad90-4aa3251849d1 service nova] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Received event network-vif-deleted-691b9a22-1a0a-4a84-a136-b828743aff9c {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 614.454105] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.531493] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.531883] env[62914]: DEBUG nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 614.534590] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.247s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.559779] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352392, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056945} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.559779] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 614.561847] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6f4d93-46dd-4dcb-8a0c-8d9525456316 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.582764] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Reconfiguring VM instance instance-00000013 to attach disk [datastore1] f3509973-013e-45ce-87f9-357e782f26d7/f3509973-013e-45ce-87f9-357e782f26d7.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 614.584094] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7000ddf-80b7-4ad7-94bf-405aceb7fa51 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.603728] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 614.603728] env[62914]: value = "task-1352393" [ 614.603728] env[62914]: _type = "Task" [ 614.603728] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.612230] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352393, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.047019] env[62914]: DEBUG nova.compute.utils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.047019] env[62914]: DEBUG nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 615.047019] env[62914]: DEBUG nova.network.neutron [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 615.116144] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352393, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.125897] env[62914]: DEBUG nova.policy [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b45ba0ad5a834147a064b83b7730fde9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '455c8a5ff8c24e30a74432bed7a7fd54', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 615.187329] env[62914]: INFO nova.scheduler.client.report [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Deleted allocations for instance 723105bc-6810-4759-8240-af0a128475e4 [ 615.550971] env[62914]: DEBUG nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 615.618572] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352393, 'name': ReconfigVM_Task, 'duration_secs': 0.749612} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.618860] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Reconfigured VM instance instance-00000013 to attach disk [datastore1] f3509973-013e-45ce-87f9-357e782f26d7/f3509973-013e-45ce-87f9-357e782f26d7.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 615.619474] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-085dfc24-b4f7-4b64-a846-a88c346e6c62 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.625030] env[62914]: DEBUG nova.network.neutron [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Successfully created port: c886397b-a988-45ab-9cdc-21a3afc92a24 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 615.632574] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 615.632574] env[62914]: value = "task-1352394" [ 615.632574] env[62914]: _type = "Task" [ 615.632574] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.641154] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352394, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.643629] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654cbdc9-d0f1-4e80-8519-70db9e2a0064 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.651150] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694e343e-68cb-41aa-85bc-e6d28da6f70a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.680922] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918fef03-cbfb-473c-86e6-2946b4f98c51 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.689667] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7948c112-84c5-4934-b6cb-a063fc82dbf5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.703020] env[62914]: DEBUG nova.compute.provider_tree [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.704952] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3ae9f6b9-a3b1-4350-8b7e-f631393f0d65 tempest-ServersNegativeTestJSON-1492438019 tempest-ServersNegativeTestJSON-1492438019-project-member] Lock "723105bc-6810-4759-8240-af0a128475e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.603s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.150358] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352394, 'name': Rename_Task, 'duration_secs': 0.133198} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.150711] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 616.150975] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10aa5299-f60c-4d8b-9e0a-a477d10648e5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.158142] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 616.158142] env[62914]: value = "task-1352395" [ 616.158142] env[62914]: _type = "Task" [ 616.158142] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.173147] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352395, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.207171] env[62914]: DEBUG nova.scheduler.client.report [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.212237] env[62914]: DEBUG nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 616.561100] env[62914]: DEBUG nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 616.600246] env[62914]: DEBUG nova.virt.hardware [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 616.600551] env[62914]: DEBUG nova.virt.hardware [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 616.600683] env[62914]: DEBUG nova.virt.hardware [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.600830] env[62914]: DEBUG nova.virt.hardware [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 616.600990] env[62914]: DEBUG nova.virt.hardware [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.601112] env[62914]: DEBUG nova.virt.hardware [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 616.601313] env[62914]: DEBUG nova.virt.hardware [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 616.601487] env[62914]: DEBUG nova.virt.hardware [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 616.601648] env[62914]: DEBUG nova.virt.hardware [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 616.601846] env[62914]: DEBUG nova.virt.hardware [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 616.602065] env[62914]: DEBUG nova.virt.hardware [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.602956] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf5669e-5ea5-482e-bd9e-c2e293e64789 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.614596] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08219dd-b2a0-4a66-9c7e-5ef6cc5cba73 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.669389] env[62914]: DEBUG oslo_vmware.api [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352395, 'name': PowerOnVM_Task, 'duration_secs': 0.440037} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.669787] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 616.670521] env[62914]: DEBUG nova.compute.manager [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 616.671858] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2287621-8f91-44f7-a4f4-bba64535d37b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.715022] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.177s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.715022] env[62914]: ERROR nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b81867b2-0504-426f-9262-65caf32e66fd, please check neutron logs for more information. [ 616.715022] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Traceback (most recent call last): [ 616.715022] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 616.715022] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] self.driver.spawn(context, instance, image_meta, [ 616.715022] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 616.715022] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.715022] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.715022] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] vm_ref = self.build_virtual_machine(instance, [ 616.715326] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.715326] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.715326] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.715326] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] for vif in network_info: [ 616.715326] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.715326] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] return self._sync_wrapper(fn, *args, **kwargs) [ 616.715326] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.715326] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] self.wait() [ 616.715326] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.715326] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] self[:] = self._gt.wait() [ 616.715326] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.715326] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] return self._exit_event.wait() [ 616.715326] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 616.715647] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] result = hub.switch() [ 616.715647] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 616.715647] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] return self.greenlet.switch() [ 616.715647] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.715647] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] result = function(*args, **kwargs) [ 616.715647] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 616.715647] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] return func(*args, **kwargs) [ 616.715647] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.715647] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] raise e [ 616.715647] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.715647] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] nwinfo = self.network_api.allocate_for_instance( [ 616.715647] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 616.715647] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] created_port_ids = self._update_ports_for_instance( [ 616.715937] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 616.715937] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] with excutils.save_and_reraise_exception(): [ 616.715937] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.715937] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] self.force_reraise() [ 616.715937] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.715937] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] raise self.value [ 616.715937] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 616.715937] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] updated_port = self._update_port( [ 616.715937] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.715937] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] _ensure_no_port_binding_failure(port) [ 616.715937] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.715937] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] raise exception.PortBindingFailed(port_id=port['id']) [ 616.716217] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] nova.exception.PortBindingFailed: Binding failed for port b81867b2-0504-426f-9262-65caf32e66fd, please check neutron logs for more information. [ 616.716217] env[62914]: ERROR nova.compute.manager [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] [ 616.716217] env[62914]: DEBUG nova.compute.utils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Binding failed for port b81867b2-0504-426f-9262-65caf32e66fd, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 616.716217] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.940s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.718590] env[62914]: INFO nova.compute.claims [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 616.723047] env[62914]: DEBUG nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Build of instance bfcba422-5ed0-4bfd-83e9-cdde324d899f was re-scheduled: Binding failed for port b81867b2-0504-426f-9262-65caf32e66fd, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 616.723940] env[62914]: DEBUG nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 616.724523] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Acquiring lock "refresh_cache-bfcba422-5ed0-4bfd-83e9-cdde324d899f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.724765] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Acquired lock "refresh_cache-bfcba422-5ed0-4bfd-83e9-cdde324d899f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.724994] env[62914]: DEBUG nova.network.neutron [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 616.755078] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.855153] env[62914]: DEBUG nova.compute.manager [req-44248d7e-0e64-49f2-8394-0995dd0ca1d0 req-1df97940-2c07-4674-9c1d-fada055989c2 service nova] [instance: 964f98e4-3462-4362-b551-576302776f3a] Received event network-changed-c886397b-a988-45ab-9cdc-21a3afc92a24 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 616.855153] env[62914]: DEBUG nova.compute.manager [req-44248d7e-0e64-49f2-8394-0995dd0ca1d0 req-1df97940-2c07-4674-9c1d-fada055989c2 service nova] [instance: 964f98e4-3462-4362-b551-576302776f3a] Refreshing instance network info cache due to event network-changed-c886397b-a988-45ab-9cdc-21a3afc92a24. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 616.855153] env[62914]: DEBUG oslo_concurrency.lockutils [req-44248d7e-0e64-49f2-8394-0995dd0ca1d0 req-1df97940-2c07-4674-9c1d-fada055989c2 service nova] Acquiring lock "refresh_cache-964f98e4-3462-4362-b551-576302776f3a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.855153] env[62914]: DEBUG oslo_concurrency.lockutils [req-44248d7e-0e64-49f2-8394-0995dd0ca1d0 req-1df97940-2c07-4674-9c1d-fada055989c2 service nova] Acquired lock "refresh_cache-964f98e4-3462-4362-b551-576302776f3a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.855153] env[62914]: DEBUG nova.network.neutron [req-44248d7e-0e64-49f2-8394-0995dd0ca1d0 req-1df97940-2c07-4674-9c1d-fada055989c2 service nova] [instance: 964f98e4-3462-4362-b551-576302776f3a] Refreshing network info cache for port c886397b-a988-45ab-9cdc-21a3afc92a24 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 617.190280] env[62914]: DEBUG oslo_concurrency.lockutils [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.248779] env[62914]: DEBUG nova.network.neutron [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.327017] env[62914]: DEBUG nova.network.neutron [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.375748] env[62914]: DEBUG nova.network.neutron [req-44248d7e-0e64-49f2-8394-0995dd0ca1d0 req-1df97940-2c07-4674-9c1d-fada055989c2 service nova] [instance: 964f98e4-3462-4362-b551-576302776f3a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.414328] env[62914]: ERROR nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c886397b-a988-45ab-9cdc-21a3afc92a24, please check neutron logs for more information. [ 617.414328] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 617.414328] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.414328] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 617.414328] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.414328] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 617.414328] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.414328] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 617.414328] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.414328] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 617.414328] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.414328] env[62914]: ERROR nova.compute.manager raise self.value [ 617.414328] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.414328] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 617.414328] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.414328] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 617.414779] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.414779] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 617.414779] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c886397b-a988-45ab-9cdc-21a3afc92a24, please check neutron logs for more information. [ 617.414779] env[62914]: ERROR nova.compute.manager [ 617.414779] env[62914]: Traceback (most recent call last): [ 617.414779] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 617.414779] env[62914]: listener.cb(fileno) [ 617.414779] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.414779] env[62914]: result = function(*args, **kwargs) [ 617.414779] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 617.414779] env[62914]: return func(*args, **kwargs) [ 617.414779] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.414779] env[62914]: raise e [ 617.414779] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.414779] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 617.414779] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.414779] env[62914]: created_port_ids = self._update_ports_for_instance( [ 617.414779] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.414779] env[62914]: with excutils.save_and_reraise_exception(): [ 617.414779] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.414779] env[62914]: self.force_reraise() [ 617.414779] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.414779] env[62914]: raise self.value [ 617.414779] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.414779] env[62914]: updated_port = self._update_port( [ 617.414779] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.414779] env[62914]: _ensure_no_port_binding_failure(port) [ 617.414779] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.414779] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 617.415542] env[62914]: nova.exception.PortBindingFailed: Binding failed for port c886397b-a988-45ab-9cdc-21a3afc92a24, please check neutron logs for more information. [ 617.415542] env[62914]: Removing descriptor: 19 [ 617.415542] env[62914]: ERROR nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c886397b-a988-45ab-9cdc-21a3afc92a24, please check neutron logs for more information. [ 617.415542] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] Traceback (most recent call last): [ 617.415542] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 617.415542] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] yield resources [ 617.415542] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 617.415542] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] self.driver.spawn(context, instance, image_meta, [ 617.415542] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 617.415542] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.415542] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.415542] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] vm_ref = self.build_virtual_machine(instance, [ 617.415900] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.415900] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.415900] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.415900] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] for vif in network_info: [ 617.415900] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 617.415900] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] return self._sync_wrapper(fn, *args, **kwargs) [ 617.415900] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 617.415900] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] self.wait() [ 617.415900] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 617.415900] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] self[:] = self._gt.wait() [ 617.415900] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.415900] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] return self._exit_event.wait() [ 617.415900] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.416328] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] result = hub.switch() [ 617.416328] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.416328] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] return self.greenlet.switch() [ 617.416328] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.416328] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] result = function(*args, **kwargs) [ 617.416328] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 617.416328] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] return func(*args, **kwargs) [ 617.416328] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.416328] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] raise e [ 617.416328] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.416328] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] nwinfo = self.network_api.allocate_for_instance( [ 617.416328] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.416328] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] created_port_ids = self._update_ports_for_instance( [ 617.416736] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.416736] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] with excutils.save_and_reraise_exception(): [ 617.416736] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.416736] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] self.force_reraise() [ 617.416736] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.416736] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] raise self.value [ 617.416736] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.416736] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] updated_port = self._update_port( [ 617.416736] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.416736] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] _ensure_no_port_binding_failure(port) [ 617.416736] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.416736] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] raise exception.PortBindingFailed(port_id=port['id']) [ 617.417084] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] nova.exception.PortBindingFailed: Binding failed for port c886397b-a988-45ab-9cdc-21a3afc92a24, please check neutron logs for more information. [ 617.417084] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] [ 617.417084] env[62914]: INFO nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Terminating instance [ 617.421871] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Acquiring lock "refresh_cache-964f98e4-3462-4362-b551-576302776f3a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.457572] env[62914]: DEBUG nova.network.neutron [req-44248d7e-0e64-49f2-8394-0995dd0ca1d0 req-1df97940-2c07-4674-9c1d-fada055989c2 service nova] [instance: 964f98e4-3462-4362-b551-576302776f3a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.828749] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Releasing lock "refresh_cache-bfcba422-5ed0-4bfd-83e9-cdde324d899f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.829027] env[62914]: DEBUG nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 617.829374] env[62914]: DEBUG nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 617.829374] env[62914]: DEBUG nova.network.neutron [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 617.850365] env[62914]: DEBUG nova.network.neutron [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.967056] env[62914]: DEBUG oslo_concurrency.lockutils [req-44248d7e-0e64-49f2-8394-0995dd0ca1d0 req-1df97940-2c07-4674-9c1d-fada055989c2 service nova] Releasing lock "refresh_cache-964f98e4-3462-4362-b551-576302776f3a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.967455] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Acquired lock "refresh_cache-964f98e4-3462-4362-b551-576302776f3a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.967642] env[62914]: DEBUG nova.network.neutron [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 618.122726] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "364c5d87-f261-439a-920e-874ee0bcf83b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.125974] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "364c5d87-f261-439a-920e-874ee0bcf83b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.245468] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb86ea50-aeda-4673-891d-e0ddf0c26319 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.256022] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3bf9858-dc07-458f-985a-8b0c13782df3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.286275] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dcbd95f-7966-4512-b729-fa8ecf033429 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.294131] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd90f96-3dbc-48c0-b43d-33e3772a993e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.307426] env[62914]: DEBUG nova.compute.provider_tree [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.312248] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "f3509973-013e-45ce-87f9-357e782f26d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.312471] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "f3509973-013e-45ce-87f9-357e782f26d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.312685] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "f3509973-013e-45ce-87f9-357e782f26d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.312972] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "f3509973-013e-45ce-87f9-357e782f26d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.313055] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "f3509973-013e-45ce-87f9-357e782f26d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.314954] env[62914]: INFO nova.compute.manager [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Terminating instance [ 618.316878] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "refresh_cache-f3509973-013e-45ce-87f9-357e782f26d7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.317045] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquired lock "refresh_cache-f3509973-013e-45ce-87f9-357e782f26d7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.317208] env[62914]: DEBUG nova.network.neutron [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 618.353199] env[62914]: DEBUG nova.network.neutron [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.494212] env[62914]: DEBUG nova.network.neutron [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.619588] env[62914]: DEBUG nova.network.neutron [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.811560] env[62914]: DEBUG nova.scheduler.client.report [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.834491] env[62914]: DEBUG nova.network.neutron [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.856130] env[62914]: INFO nova.compute.manager [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] [instance: bfcba422-5ed0-4bfd-83e9-cdde324d899f] Took 1.03 seconds to deallocate network for instance. [ 618.877657] env[62914]: DEBUG nova.compute.manager [req-0ac8383b-a46b-4491-a509-f29c5ec8030b req-64809940-0309-44d9-9c92-2f40fbd547e4 service nova] [instance: 964f98e4-3462-4362-b551-576302776f3a] Received event network-vif-deleted-c886397b-a988-45ab-9cdc-21a3afc92a24 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.895756] env[62914]: DEBUG nova.network.neutron [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.122339] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Releasing lock "refresh_cache-964f98e4-3462-4362-b551-576302776f3a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.122842] env[62914]: DEBUG nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 619.123103] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 619.123463] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0fc99802-f982-40d9-bf28-5d7a3c70bba4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.133557] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97272d65-e61c-4c91-a5a8-01efc4b54ed1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.158350] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 964f98e4-3462-4362-b551-576302776f3a could not be found. [ 619.158569] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 619.158753] env[62914]: INFO nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 619.159243] env[62914]: DEBUG oslo.service.loopingcall [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 619.159243] env[62914]: DEBUG nova.compute.manager [-] [instance: 964f98e4-3462-4362-b551-576302776f3a] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 619.159320] env[62914]: DEBUG nova.network.neutron [-] [instance: 964f98e4-3462-4362-b551-576302776f3a] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 619.176337] env[62914]: DEBUG nova.network.neutron [-] [instance: 964f98e4-3462-4362-b551-576302776f3a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.317533] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.317819] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 619.320378] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.480s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.321763] env[62914]: INFO nova.compute.claims [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.398155] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Releasing lock "refresh_cache-f3509973-013e-45ce-87f9-357e782f26d7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.398544] env[62914]: DEBUG nova.compute.manager [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 619.398734] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 619.399814] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85b0732-914a-48d8-81a8-a380fd5e5abf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.408020] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 619.408292] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9143ab2-d8d6-4e33-9c52-9084a5f23c49 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.414180] env[62914]: DEBUG oslo_vmware.api [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 619.414180] env[62914]: value = "task-1352396" [ 619.414180] env[62914]: _type = "Task" [ 619.414180] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.423685] env[62914]: DEBUG oslo_vmware.api [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352396, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.679246] env[62914]: DEBUG nova.network.neutron [-] [instance: 964f98e4-3462-4362-b551-576302776f3a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.825826] env[62914]: DEBUG nova.compute.utils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 619.828986] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 619.829169] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 619.871751] env[62914]: DEBUG nova.policy [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f05fcb5e69eb49759ed244e98cea0bec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd182d40cb72d4107a6902d06d6e75aa2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 619.891024] env[62914]: INFO nova.scheduler.client.report [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Deleted allocations for instance bfcba422-5ed0-4bfd-83e9-cdde324d899f [ 619.924968] env[62914]: DEBUG oslo_vmware.api [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352396, 'name': PowerOffVM_Task, 'duration_secs': 0.112703} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.925316] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 619.925514] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 619.925764] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4db0b1f4-60e1-43d5-b3b9-d96517e7ada6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.951687] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 619.951927] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 619.953233] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Deleting the datastore file [datastore1] f3509973-013e-45ce-87f9-357e782f26d7 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 619.953233] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-280a30cf-1cf6-4ce7-84f2-fd4c7f429f56 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.958790] env[62914]: DEBUG oslo_vmware.api [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 619.958790] env[62914]: value = "task-1352398" [ 619.958790] env[62914]: _type = "Task" [ 619.958790] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.967098] env[62914]: DEBUG oslo_vmware.api [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352398, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.184629] env[62914]: INFO nova.compute.manager [-] [instance: 964f98e4-3462-4362-b551-576302776f3a] Took 1.02 seconds to deallocate network for instance. [ 620.185351] env[62914]: DEBUG nova.compute.claims [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 620.185715] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.200679] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Successfully created port: b24a9394-2930-43a3-a75c-b52c67465668 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 620.334083] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 620.399380] env[62914]: DEBUG oslo_concurrency.lockutils [None req-51b2eff3-0893-4e6f-9989-9e82044f8cab tempest-AttachInterfacesV270Test-1928111885 tempest-AttachInterfacesV270Test-1928111885-project-member] Lock "bfcba422-5ed0-4bfd-83e9-cdde324d899f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.740s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.415693] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 620.415693] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 620.470475] env[62914]: DEBUG oslo_vmware.api [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352398, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.794745] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32a7767-9ca6-49c1-8a6c-7999d5b31c35 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.802211] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d851f28-36c3-4983-b1de-c283ef2e507f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.847786] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03632c1-390f-404e-96f3-c3319632767b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.856776] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c8c03d-0cc6-4bdd-9356-4484f01c0012 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.872431] env[62914]: DEBUG nova.compute.provider_tree [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.901923] env[62914]: DEBUG nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 620.924034] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 620.924217] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Starting heal instance info cache {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 620.924338] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Rebuilding the list of instances to heal {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 620.969219] env[62914]: DEBUG oslo_vmware.api [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352398, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.020240] env[62914]: DEBUG nova.compute.manager [req-f8d96c13-db7e-4213-a611-8787c6a264d7 req-0eb5e201-5700-45a9-89c1-b3c44fb12651 service nova] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Received event network-changed-b24a9394-2930-43a3-a75c-b52c67465668 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 621.020443] env[62914]: DEBUG nova.compute.manager [req-f8d96c13-db7e-4213-a611-8787c6a264d7 req-0eb5e201-5700-45a9-89c1-b3c44fb12651 service nova] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Refreshing instance network info cache due to event network-changed-b24a9394-2930-43a3-a75c-b52c67465668. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 621.020653] env[62914]: DEBUG oslo_concurrency.lockutils [req-f8d96c13-db7e-4213-a611-8787c6a264d7 req-0eb5e201-5700-45a9-89c1-b3c44fb12651 service nova] Acquiring lock "refresh_cache-283ffd74-3891-4b5e-bcf5-49b265978051" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.020789] env[62914]: DEBUG oslo_concurrency.lockutils [req-f8d96c13-db7e-4213-a611-8787c6a264d7 req-0eb5e201-5700-45a9-89c1-b3c44fb12651 service nova] Acquired lock "refresh_cache-283ffd74-3891-4b5e-bcf5-49b265978051" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.020949] env[62914]: DEBUG nova.network.neutron [req-f8d96c13-db7e-4213-a611-8787c6a264d7 req-0eb5e201-5700-45a9-89c1-b3c44fb12651 service nova] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Refreshing network info cache for port b24a9394-2930-43a3-a75c-b52c67465668 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 621.165155] env[62914]: ERROR nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b24a9394-2930-43a3-a75c-b52c67465668, please check neutron logs for more information. [ 621.165155] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 621.165155] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.165155] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 621.165155] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.165155] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 621.165155] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.165155] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 621.165155] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.165155] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 621.165155] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.165155] env[62914]: ERROR nova.compute.manager raise self.value [ 621.165155] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.165155] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 621.165155] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.165155] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 621.165722] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.165722] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 621.165722] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b24a9394-2930-43a3-a75c-b52c67465668, please check neutron logs for more information. [ 621.165722] env[62914]: ERROR nova.compute.manager [ 621.165722] env[62914]: Traceback (most recent call last): [ 621.165722] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 621.165722] env[62914]: listener.cb(fileno) [ 621.165722] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.165722] env[62914]: result = function(*args, **kwargs) [ 621.165722] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.165722] env[62914]: return func(*args, **kwargs) [ 621.165722] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.165722] env[62914]: raise e [ 621.165722] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.165722] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 621.165722] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.165722] env[62914]: created_port_ids = self._update_ports_for_instance( [ 621.165722] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.165722] env[62914]: with excutils.save_and_reraise_exception(): [ 621.165722] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.165722] env[62914]: self.force_reraise() [ 621.165722] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.165722] env[62914]: raise self.value [ 621.165722] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.165722] env[62914]: updated_port = self._update_port( [ 621.165722] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.165722] env[62914]: _ensure_no_port_binding_failure(port) [ 621.165722] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.165722] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 621.166883] env[62914]: nova.exception.PortBindingFailed: Binding failed for port b24a9394-2930-43a3-a75c-b52c67465668, please check neutron logs for more information. [ 621.166883] env[62914]: Removing descriptor: 15 [ 621.434542] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 621.436809] env[62914]: DEBUG nova.scheduler.client.report [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.449228] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Skipping network cache update for instance because it is being deleted. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 621.449228] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 621.449228] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 621.449228] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 964f98e4-3462-4362-b551-576302776f3a] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 621.449228] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 621.449228] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 621.480669] env[62914]: DEBUG oslo_vmware.api [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352398, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.483912] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.487545] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "refresh_cache-34b0b81e-fc98-4019-92c3-fe2ce2fdfd67" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.487677] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquired lock "refresh_cache-34b0b81e-fc98-4019-92c3-fe2ce2fdfd67" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.489480] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Forcefully refreshing network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 621.489679] env[62914]: DEBUG nova.objects.instance [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lazy-loading 'info_cache' on Instance uuid 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 621.493167] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 621.493408] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 621.493562] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.493739] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 621.494341] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.494341] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 621.494341] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 621.495955] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 621.495955] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 621.495955] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 621.495955] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.496187] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19badb1-a00f-4e59-9b4a-2e3d6d2eeed9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.506438] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b0cbc7-bfc5-4208-8eb4-03d2acb06290 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.522587] env[62914]: ERROR nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b24a9394-2930-43a3-a75c-b52c67465668, please check neutron logs for more information. [ 621.522587] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Traceback (most recent call last): [ 621.522587] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 621.522587] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] yield resources [ 621.522587] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.522587] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] self.driver.spawn(context, instance, image_meta, [ 621.522587] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 621.522587] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.522587] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.522587] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] vm_ref = self.build_virtual_machine(instance, [ 621.522587] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.523246] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.523246] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.523246] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] for vif in network_info: [ 621.523246] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.523246] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] return self._sync_wrapper(fn, *args, **kwargs) [ 621.523246] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.523246] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] self.wait() [ 621.523246] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.523246] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] self[:] = self._gt.wait() [ 621.523246] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.523246] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] return self._exit_event.wait() [ 621.523246] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 621.523246] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] current.throw(*self._exc) [ 621.523770] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.523770] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] result = function(*args, **kwargs) [ 621.523770] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.523770] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] return func(*args, **kwargs) [ 621.523770] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.523770] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] raise e [ 621.523770] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.523770] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] nwinfo = self.network_api.allocate_for_instance( [ 621.523770] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.523770] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] created_port_ids = self._update_ports_for_instance( [ 621.523770] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.523770] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] with excutils.save_and_reraise_exception(): [ 621.523770] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.524453] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] self.force_reraise() [ 621.524453] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.524453] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] raise self.value [ 621.524453] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.524453] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] updated_port = self._update_port( [ 621.524453] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.524453] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] _ensure_no_port_binding_failure(port) [ 621.524453] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.524453] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] raise exception.PortBindingFailed(port_id=port['id']) [ 621.524453] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] nova.exception.PortBindingFailed: Binding failed for port b24a9394-2930-43a3-a75c-b52c67465668, please check neutron logs for more information. [ 621.524453] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] [ 621.524453] env[62914]: INFO nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Terminating instance [ 621.527169] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "refresh_cache-283ffd74-3891-4b5e-bcf5-49b265978051" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.549103] env[62914]: DEBUG nova.network.neutron [req-f8d96c13-db7e-4213-a611-8787c6a264d7 req-0eb5e201-5700-45a9-89c1-b3c44fb12651 service nova] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.705013] env[62914]: DEBUG nova.network.neutron [req-f8d96c13-db7e-4213-a611-8787c6a264d7 req-0eb5e201-5700-45a9-89c1-b3c44fb12651 service nova] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.954344] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.634s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.954896] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 621.958877] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.527s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.960638] env[62914]: INFO nova.compute.claims [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.971639] env[62914]: DEBUG oslo_vmware.api [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352398, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.664224} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.971880] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 621.972064] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 621.972243] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 621.972412] env[62914]: INFO nova.compute.manager [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Took 2.57 seconds to destroy the instance on the hypervisor. [ 621.972644] env[62914]: DEBUG oslo.service.loopingcall [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 621.973139] env[62914]: DEBUG nova.compute.manager [-] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 621.973241] env[62914]: DEBUG nova.network.neutron [-] [instance: f3509973-013e-45ce-87f9-357e782f26d7] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 621.990997] env[62914]: DEBUG nova.network.neutron [-] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.208978] env[62914]: DEBUG oslo_concurrency.lockutils [req-f8d96c13-db7e-4213-a611-8787c6a264d7 req-0eb5e201-5700-45a9-89c1-b3c44fb12651 service nova] Releasing lock "refresh_cache-283ffd74-3891-4b5e-bcf5-49b265978051" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.208978] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquired lock "refresh_cache-283ffd74-3891-4b5e-bcf5-49b265978051" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.208978] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 622.465476] env[62914]: DEBUG nova.compute.utils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 622.470415] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 622.470415] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 622.495333] env[62914]: DEBUG nova.network.neutron [-] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.542085] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.554036] env[62914]: DEBUG nova.policy [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f05fcb5e69eb49759ed244e98cea0bec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd182d40cb72d4107a6902d06d6e75aa2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 622.728348] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.825763] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.970384] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 622.998078] env[62914]: INFO nova.compute.manager [-] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Took 1.02 seconds to deallocate network for instance. [ 623.056393] env[62914]: DEBUG nova.compute.manager [req-83cd6b7c-7847-4fec-9f0f-2c9308be1e98 req-8b368a27-f07e-494c-b475-3105bdb93a49 service nova] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Received event network-vif-deleted-b24a9394-2930-43a3-a75c-b52c67465668 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 623.072654] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Successfully created port: a3d89c65-4294-4a4e-b974-ed7f91fc188a {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.148710] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.328931] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Releasing lock "refresh_cache-283ffd74-3891-4b5e-bcf5-49b265978051" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.329346] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 623.329535] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 623.331928] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d1da687e-9c0f-435c-824f-05bdd5611c98 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.346648] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef422429-3deb-44a2-85c5-9737564ce0f7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.375365] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 283ffd74-3891-4b5e-bcf5-49b265978051 could not be found. [ 623.375629] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 623.375812] env[62914]: INFO nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Took 0.05 seconds to destroy the instance on the hypervisor. [ 623.376079] env[62914]: DEBUG oslo.service.loopingcall [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 623.378401] env[62914]: DEBUG nova.compute.manager [-] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.378500] env[62914]: DEBUG nova.network.neutron [-] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 623.397901] env[62914]: DEBUG nova.network.neutron [-] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.486811] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f583818-906a-43c7-aa91-859ff3916b89 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.494574] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d259aa94-7a85-41ee-87a4-d01130a23e3a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.525877] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.526932] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2399ccd0-3f0d-422f-bd03-51d1ea688411 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.534180] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8ad750-8f6a-4e3d-b006-48eeb34cf97b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.547869] env[62914]: DEBUG nova.compute.provider_tree [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.651080] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Releasing lock "refresh_cache-34b0b81e-fc98-4019-92c3-fe2ce2fdfd67" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.651244] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Updated the network info_cache for instance {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 623.651459] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 623.651615] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 623.651767] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 623.651898] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 623.652044] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 623.652186] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 623.652859] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62914) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 623.652859] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager.update_available_resource {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 623.900625] env[62914]: DEBUG nova.network.neutron [-] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.991314] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 624.022101] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 624.022510] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 624.022510] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 624.022684] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 624.023098] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 624.023098] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 624.023210] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 624.023716] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 624.023716] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 624.023716] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 624.023906] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 624.024689] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05bd98d0-7610-45b8-8552-bdc1cbc5e51c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.037671] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08c3f18-6eef-4e2d-82e5-c3d61e1d60d5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.052129] env[62914]: DEBUG nova.scheduler.client.report [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.074488] env[62914]: ERROR nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a3d89c65-4294-4a4e-b974-ed7f91fc188a, please check neutron logs for more information. [ 624.074488] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 624.074488] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.074488] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 624.074488] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.074488] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 624.074488] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.074488] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 624.074488] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.074488] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 624.074488] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.074488] env[62914]: ERROR nova.compute.manager raise self.value [ 624.074488] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.074488] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 624.074488] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.074488] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 624.075221] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.075221] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 624.075221] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a3d89c65-4294-4a4e-b974-ed7f91fc188a, please check neutron logs for more information. [ 624.075221] env[62914]: ERROR nova.compute.manager [ 624.075221] env[62914]: Traceback (most recent call last): [ 624.075221] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 624.075221] env[62914]: listener.cb(fileno) [ 624.075221] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.075221] env[62914]: result = function(*args, **kwargs) [ 624.075221] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.075221] env[62914]: return func(*args, **kwargs) [ 624.075221] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.075221] env[62914]: raise e [ 624.075221] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.075221] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 624.075221] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.075221] env[62914]: created_port_ids = self._update_ports_for_instance( [ 624.075221] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.075221] env[62914]: with excutils.save_and_reraise_exception(): [ 624.075221] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.075221] env[62914]: self.force_reraise() [ 624.075221] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.075221] env[62914]: raise self.value [ 624.075221] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.075221] env[62914]: updated_port = self._update_port( [ 624.075221] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.075221] env[62914]: _ensure_no_port_binding_failure(port) [ 624.075221] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.075221] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 624.075939] env[62914]: nova.exception.PortBindingFailed: Binding failed for port a3d89c65-4294-4a4e-b974-ed7f91fc188a, please check neutron logs for more information. [ 624.075939] env[62914]: Removing descriptor: 15 [ 624.075939] env[62914]: ERROR nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a3d89c65-4294-4a4e-b974-ed7f91fc188a, please check neutron logs for more information. [ 624.075939] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Traceback (most recent call last): [ 624.075939] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 624.075939] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] yield resources [ 624.075939] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.075939] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] self.driver.spawn(context, instance, image_meta, [ 624.075939] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 624.075939] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.075939] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.075939] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] vm_ref = self.build_virtual_machine(instance, [ 624.076283] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.076283] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.076283] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.076283] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] for vif in network_info: [ 624.076283] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.076283] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] return self._sync_wrapper(fn, *args, **kwargs) [ 624.076283] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.076283] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] self.wait() [ 624.076283] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.076283] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] self[:] = self._gt.wait() [ 624.076283] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.076283] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] return self._exit_event.wait() [ 624.076283] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.077651] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] result = hub.switch() [ 624.077651] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.077651] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] return self.greenlet.switch() [ 624.077651] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.077651] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] result = function(*args, **kwargs) [ 624.077651] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.077651] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] return func(*args, **kwargs) [ 624.077651] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.077651] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] raise e [ 624.077651] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.077651] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] nwinfo = self.network_api.allocate_for_instance( [ 624.077651] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.077651] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] created_port_ids = self._update_ports_for_instance( [ 624.078045] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.078045] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] with excutils.save_and_reraise_exception(): [ 624.078045] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.078045] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] self.force_reraise() [ 624.078045] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.078045] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] raise self.value [ 624.078045] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.078045] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] updated_port = self._update_port( [ 624.078045] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.078045] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] _ensure_no_port_binding_failure(port) [ 624.078045] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.078045] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] raise exception.PortBindingFailed(port_id=port['id']) [ 624.078333] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] nova.exception.PortBindingFailed: Binding failed for port a3d89c65-4294-4a4e-b974-ed7f91fc188a, please check neutron logs for more information. [ 624.078333] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] [ 624.078333] env[62914]: INFO nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Terminating instance [ 624.079524] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "refresh_cache-bfcba5cb-d9c4-4ddd-9018-d316c13525d9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.079691] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquired lock "refresh_cache-bfcba5cb-d9c4-4ddd-9018-d316c13525d9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.079868] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.155020] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.404607] env[62914]: INFO nova.compute.manager [-] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Took 1.03 seconds to deallocate network for instance. [ 624.407425] env[62914]: DEBUG nova.compute.claims [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 624.407611] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.557332] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.598s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.557864] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 624.560502] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.419s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.561856] env[62914]: INFO nova.compute.claims [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.602550] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.697020] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.066950] env[62914]: DEBUG nova.compute.utils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 625.074129] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 625.074129] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 625.106276] env[62914]: DEBUG nova.compute.manager [req-6792d672-ff34-4648-bbc0-fc7d8ce0da1c req-0ff2ae5e-54cf-491f-ab71-3f7eed966958 service nova] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Received event network-changed-a3d89c65-4294-4a4e-b974-ed7f91fc188a {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 625.106496] env[62914]: DEBUG nova.compute.manager [req-6792d672-ff34-4648-bbc0-fc7d8ce0da1c req-0ff2ae5e-54cf-491f-ab71-3f7eed966958 service nova] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Refreshing instance network info cache due to event network-changed-a3d89c65-4294-4a4e-b974-ed7f91fc188a. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 625.106697] env[62914]: DEBUG oslo_concurrency.lockutils [req-6792d672-ff34-4648-bbc0-fc7d8ce0da1c req-0ff2ae5e-54cf-491f-ab71-3f7eed966958 service nova] Acquiring lock "refresh_cache-bfcba5cb-d9c4-4ddd-9018-d316c13525d9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.114445] env[62914]: DEBUG nova.policy [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f05fcb5e69eb49759ed244e98cea0bec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd182d40cb72d4107a6902d06d6e75aa2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 625.200200] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Releasing lock "refresh_cache-bfcba5cb-d9c4-4ddd-9018-d316c13525d9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.200386] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 625.200598] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 625.200930] env[62914]: DEBUG oslo_concurrency.lockutils [req-6792d672-ff34-4648-bbc0-fc7d8ce0da1c req-0ff2ae5e-54cf-491f-ab71-3f7eed966958 service nova] Acquired lock "refresh_cache-bfcba5cb-d9c4-4ddd-9018-d316c13525d9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.201128] env[62914]: DEBUG nova.network.neutron [req-6792d672-ff34-4648-bbc0-fc7d8ce0da1c req-0ff2ae5e-54cf-491f-ab71-3f7eed966958 service nova] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Refreshing network info cache for port a3d89c65-4294-4a4e-b974-ed7f91fc188a {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 625.202381] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06742748-fdfd-483d-83df-7c3ffed4d14f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.211988] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c435df5-9e6c-480d-8ff8-b32bdbe53d1a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.241893] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bfcba5cb-d9c4-4ddd-9018-d316c13525d9 could not be found. [ 625.242130] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 625.242312] env[62914]: INFO nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 625.242553] env[62914]: DEBUG oslo.service.loopingcall [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.242805] env[62914]: DEBUG nova.compute.manager [-] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.242932] env[62914]: DEBUG nova.network.neutron [-] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.262141] env[62914]: DEBUG nova.network.neutron [-] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.415793] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Successfully created port: d730aaf3-0ed0-4e80-8eee-4257ababfe57 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 625.571408] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 625.732112] env[62914]: DEBUG nova.network.neutron [req-6792d672-ff34-4648-bbc0-fc7d8ce0da1c req-0ff2ae5e-54cf-491f-ab71-3f7eed966958 service nova] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.767406] env[62914]: DEBUG nova.network.neutron [-] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.028924] env[62914]: DEBUG nova.network.neutron [req-6792d672-ff34-4648-bbc0-fc7d8ce0da1c req-0ff2ae5e-54cf-491f-ab71-3f7eed966958 service nova] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.096872] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7a04d0-9ed8-4648-9331-b038194372f4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.104914] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab360e4-36bc-41ea-b205-ccccabd9b313 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.137519] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d44dc85-f3fc-48df-bf30-4c9033f0fecb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.146539] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e543fbe-7db3-4404-a55a-cc3a4c8dfc1f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.161775] env[62914]: DEBUG nova.compute.provider_tree [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.268659] env[62914]: INFO nova.compute.manager [-] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Took 1.03 seconds to deallocate network for instance. [ 626.270956] env[62914]: DEBUG nova.compute.claims [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 626.271166] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.526025] env[62914]: ERROR nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d730aaf3-0ed0-4e80-8eee-4257ababfe57, please check neutron logs for more information. [ 626.526025] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 626.526025] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.526025] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 626.526025] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.526025] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 626.526025] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.526025] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 626.526025] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.526025] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 626.526025] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.526025] env[62914]: ERROR nova.compute.manager raise self.value [ 626.526025] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.526025] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 626.526025] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.526025] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 626.526500] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.526500] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 626.526500] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d730aaf3-0ed0-4e80-8eee-4257ababfe57, please check neutron logs for more information. [ 626.526500] env[62914]: ERROR nova.compute.manager [ 626.526500] env[62914]: Traceback (most recent call last): [ 626.526500] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 626.526500] env[62914]: listener.cb(fileno) [ 626.526500] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.526500] env[62914]: result = function(*args, **kwargs) [ 626.526500] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.526500] env[62914]: return func(*args, **kwargs) [ 626.526500] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.526500] env[62914]: raise e [ 626.526500] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.526500] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 626.526500] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.526500] env[62914]: created_port_ids = self._update_ports_for_instance( [ 626.526500] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.526500] env[62914]: with excutils.save_and_reraise_exception(): [ 626.526500] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.526500] env[62914]: self.force_reraise() [ 626.526500] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.526500] env[62914]: raise self.value [ 626.526500] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.526500] env[62914]: updated_port = self._update_port( [ 626.526500] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.526500] env[62914]: _ensure_no_port_binding_failure(port) [ 626.526500] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.526500] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 626.527255] env[62914]: nova.exception.PortBindingFailed: Binding failed for port d730aaf3-0ed0-4e80-8eee-4257ababfe57, please check neutron logs for more information. [ 626.527255] env[62914]: Removing descriptor: 15 [ 626.532197] env[62914]: DEBUG oslo_concurrency.lockutils [req-6792d672-ff34-4648-bbc0-fc7d8ce0da1c req-0ff2ae5e-54cf-491f-ab71-3f7eed966958 service nova] Releasing lock "refresh_cache-bfcba5cb-d9c4-4ddd-9018-d316c13525d9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.532457] env[62914]: DEBUG nova.compute.manager [req-6792d672-ff34-4648-bbc0-fc7d8ce0da1c req-0ff2ae5e-54cf-491f-ab71-3f7eed966958 service nova] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Received event network-vif-deleted-a3d89c65-4294-4a4e-b974-ed7f91fc188a {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 626.586154] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 626.620338] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.620493] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.620685] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.620805] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.620945] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.621106] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.621311] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.621467] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.621631] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.621789] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.621959] env[62914]: DEBUG nova.virt.hardware [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.622950] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e37e2caa-b975-43b7-aa00-b761023392b9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.631796] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac0cba2-db48-4add-a3ef-1f9c3e3ade9c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.647054] env[62914]: ERROR nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d730aaf3-0ed0-4e80-8eee-4257ababfe57, please check neutron logs for more information. [ 626.647054] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Traceback (most recent call last): [ 626.647054] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 626.647054] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] yield resources [ 626.647054] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.647054] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] self.driver.spawn(context, instance, image_meta, [ 626.647054] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 626.647054] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.647054] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.647054] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] vm_ref = self.build_virtual_machine(instance, [ 626.647054] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.647398] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.647398] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.647398] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] for vif in network_info: [ 626.647398] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 626.647398] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] return self._sync_wrapper(fn, *args, **kwargs) [ 626.647398] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 626.647398] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] self.wait() [ 626.647398] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 626.647398] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] self[:] = self._gt.wait() [ 626.647398] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.647398] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] return self._exit_event.wait() [ 626.647398] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 626.647398] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] current.throw(*self._exc) [ 626.647701] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.647701] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] result = function(*args, **kwargs) [ 626.647701] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.647701] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] return func(*args, **kwargs) [ 626.647701] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.647701] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] raise e [ 626.647701] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.647701] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] nwinfo = self.network_api.allocate_for_instance( [ 626.647701] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.647701] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] created_port_ids = self._update_ports_for_instance( [ 626.647701] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.647701] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] with excutils.save_and_reraise_exception(): [ 626.647701] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.648022] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] self.force_reraise() [ 626.648022] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.648022] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] raise self.value [ 626.648022] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.648022] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] updated_port = self._update_port( [ 626.648022] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.648022] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] _ensure_no_port_binding_failure(port) [ 626.648022] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.648022] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] raise exception.PortBindingFailed(port_id=port['id']) [ 626.648022] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] nova.exception.PortBindingFailed: Binding failed for port d730aaf3-0ed0-4e80-8eee-4257ababfe57, please check neutron logs for more information. [ 626.648022] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] [ 626.648022] env[62914]: INFO nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Terminating instance [ 626.650384] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "refresh_cache-86784639-8ac7-4f67-be5b-05c3ab1229b7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.650384] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquired lock "refresh_cache-86784639-8ac7-4f67-be5b-05c3ab1229b7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.650384] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 626.664079] env[62914]: DEBUG nova.scheduler.client.report [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.134598] env[62914]: DEBUG nova.compute.manager [req-a2107b81-9dea-43bc-aa66-c6a5b9f66a3b req-715635ef-5b7c-4e98-ac7f-c16405d6fcf7 service nova] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Received event network-changed-d730aaf3-0ed0-4e80-8eee-4257ababfe57 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 627.134860] env[62914]: DEBUG nova.compute.manager [req-a2107b81-9dea-43bc-aa66-c6a5b9f66a3b req-715635ef-5b7c-4e98-ac7f-c16405d6fcf7 service nova] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Refreshing instance network info cache due to event network-changed-d730aaf3-0ed0-4e80-8eee-4257ababfe57. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 627.135046] env[62914]: DEBUG oslo_concurrency.lockutils [req-a2107b81-9dea-43bc-aa66-c6a5b9f66a3b req-715635ef-5b7c-4e98-ac7f-c16405d6fcf7 service nova] Acquiring lock "refresh_cache-86784639-8ac7-4f67-be5b-05c3ab1229b7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.169954] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.169954] env[62914]: DEBUG nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 627.172787] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.329s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.176018] env[62914]: INFO nova.compute.claims [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.334021] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.501849] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.678499] env[62914]: DEBUG nova.compute.utils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 627.685324] env[62914]: DEBUG nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 627.685480] env[62914]: DEBUG nova.network.neutron [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 627.753498] env[62914]: DEBUG nova.policy [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7c3e1664a6da4ee6bb53860ad83a7087', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0497bf6570de478f8b92829e031be5a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 628.005152] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Releasing lock "refresh_cache-86784639-8ac7-4f67-be5b-05c3ab1229b7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.005614] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 628.005801] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 628.006163] env[62914]: DEBUG oslo_concurrency.lockutils [req-a2107b81-9dea-43bc-aa66-c6a5b9f66a3b req-715635ef-5b7c-4e98-ac7f-c16405d6fcf7 service nova] Acquired lock "refresh_cache-86784639-8ac7-4f67-be5b-05c3ab1229b7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.006398] env[62914]: DEBUG nova.network.neutron [req-a2107b81-9dea-43bc-aa66-c6a5b9f66a3b req-715635ef-5b7c-4e98-ac7f-c16405d6fcf7 service nova] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Refreshing network info cache for port d730aaf3-0ed0-4e80-8eee-4257ababfe57 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 628.007510] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f819ad5f-1008-4b74-af02-cbcf634a35a6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.018698] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4748f939-10ab-4e6d-bb34-6b83c68f5ade {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.040882] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 86784639-8ac7-4f67-be5b-05c3ab1229b7 could not be found. [ 628.041130] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 628.041316] env[62914]: INFO nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 628.041554] env[62914]: DEBUG oslo.service.loopingcall [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 628.041786] env[62914]: DEBUG nova.compute.manager [-] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.041878] env[62914]: DEBUG nova.network.neutron [-] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.056894] env[62914]: DEBUG nova.network.neutron [-] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.189405] env[62914]: DEBUG nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 628.226709] env[62914]: DEBUG nova.network.neutron [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Successfully created port: e9e80771-8137-4491-b800-397ba5b4f512 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 628.552722] env[62914]: DEBUG nova.network.neutron [req-a2107b81-9dea-43bc-aa66-c6a5b9f66a3b req-715635ef-5b7c-4e98-ac7f-c16405d6fcf7 service nova] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.563345] env[62914]: DEBUG nova.network.neutron [-] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.732604] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6664c7-ab63-4a42-a57c-21863602c5da {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.744891] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49bfca5f-2428-42c4-8fb8-924dd2e9ca36 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.783209] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d23643e6-77e7-4121-9ba4-283279d49866 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.787633] env[62914]: DEBUG nova.network.neutron [req-a2107b81-9dea-43bc-aa66-c6a5b9f66a3b req-715635ef-5b7c-4e98-ac7f-c16405d6fcf7 service nova] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.794734] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c9a6b51-35e3-4fc5-a018-e9299d89a5a8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.810422] env[62914]: DEBUG nova.compute.provider_tree [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.069255] env[62914]: INFO nova.compute.manager [-] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Took 1.03 seconds to deallocate network for instance. [ 629.072769] env[62914]: DEBUG nova.compute.claims [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 629.072891] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.210112] env[62914]: DEBUG nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 629.248189] env[62914]: DEBUG nova.virt.hardware [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 629.248461] env[62914]: DEBUG nova.virt.hardware [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 629.248832] env[62914]: DEBUG nova.virt.hardware [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 629.249342] env[62914]: DEBUG nova.virt.hardware [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 629.249342] env[62914]: DEBUG nova.virt.hardware [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 629.249510] env[62914]: DEBUG nova.virt.hardware [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 629.249954] env[62914]: DEBUG nova.virt.hardware [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 629.250398] env[62914]: DEBUG nova.virt.hardware [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 629.250398] env[62914]: DEBUG nova.virt.hardware [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 629.250502] env[62914]: DEBUG nova.virt.hardware [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 629.250676] env[62914]: DEBUG nova.virt.hardware [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.252166] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0326375-2a19-49b7-9808-9f9c55c24b63 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.258453] env[62914]: DEBUG nova.compute.manager [req-e147aa07-c0b8-4028-b2e0-56d01bf1a1b5 req-5de146f6-c23e-4296-9263-7777c996b38a service nova] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Received event network-changed-e9e80771-8137-4491-b800-397ba5b4f512 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.258453] env[62914]: DEBUG nova.compute.manager [req-e147aa07-c0b8-4028-b2e0-56d01bf1a1b5 req-5de146f6-c23e-4296-9263-7777c996b38a service nova] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Refreshing instance network info cache due to event network-changed-e9e80771-8137-4491-b800-397ba5b4f512. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 629.259238] env[62914]: DEBUG oslo_concurrency.lockutils [req-e147aa07-c0b8-4028-b2e0-56d01bf1a1b5 req-5de146f6-c23e-4296-9263-7777c996b38a service nova] Acquiring lock "refresh_cache-8bda7658-2b86-4869-89e2-b10d7119ae18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.259426] env[62914]: DEBUG oslo_concurrency.lockutils [req-e147aa07-c0b8-4028-b2e0-56d01bf1a1b5 req-5de146f6-c23e-4296-9263-7777c996b38a service nova] Acquired lock "refresh_cache-8bda7658-2b86-4869-89e2-b10d7119ae18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.259622] env[62914]: DEBUG nova.network.neutron [req-e147aa07-c0b8-4028-b2e0-56d01bf1a1b5 req-5de146f6-c23e-4296-9263-7777c996b38a service nova] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Refreshing network info cache for port e9e80771-8137-4491-b800-397ba5b4f512 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 629.266868] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83538392-4a7d-48f4-9fad-e8d66b1a08ff {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.291786] env[62914]: DEBUG oslo_concurrency.lockutils [req-a2107b81-9dea-43bc-aa66-c6a5b9f66a3b req-715635ef-5b7c-4e98-ac7f-c16405d6fcf7 service nova] Releasing lock "refresh_cache-86784639-8ac7-4f67-be5b-05c3ab1229b7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.292060] env[62914]: DEBUG nova.compute.manager [req-a2107b81-9dea-43bc-aa66-c6a5b9f66a3b req-715635ef-5b7c-4e98-ac7f-c16405d6fcf7 service nova] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Received event network-vif-deleted-d730aaf3-0ed0-4e80-8eee-4257ababfe57 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.313193] env[62914]: DEBUG nova.scheduler.client.report [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.421943] env[62914]: ERROR nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e9e80771-8137-4491-b800-397ba5b4f512, please check neutron logs for more information. [ 629.421943] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 629.421943] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.421943] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 629.421943] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.421943] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 629.421943] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.421943] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 629.421943] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.421943] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 629.421943] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.421943] env[62914]: ERROR nova.compute.manager raise self.value [ 629.421943] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.421943] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 629.421943] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.421943] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 629.422614] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.422614] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 629.422614] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e9e80771-8137-4491-b800-397ba5b4f512, please check neutron logs for more information. [ 629.422614] env[62914]: ERROR nova.compute.manager [ 629.422614] env[62914]: Traceback (most recent call last): [ 629.422614] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 629.422614] env[62914]: listener.cb(fileno) [ 629.422614] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.422614] env[62914]: result = function(*args, **kwargs) [ 629.422614] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 629.422614] env[62914]: return func(*args, **kwargs) [ 629.422614] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.422614] env[62914]: raise e [ 629.422614] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.422614] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 629.422614] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.422614] env[62914]: created_port_ids = self._update_ports_for_instance( [ 629.422614] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.422614] env[62914]: with excutils.save_and_reraise_exception(): [ 629.422614] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.422614] env[62914]: self.force_reraise() [ 629.422614] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.422614] env[62914]: raise self.value [ 629.422614] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.422614] env[62914]: updated_port = self._update_port( [ 629.422614] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.422614] env[62914]: _ensure_no_port_binding_failure(port) [ 629.422614] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.422614] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 629.423493] env[62914]: nova.exception.PortBindingFailed: Binding failed for port e9e80771-8137-4491-b800-397ba5b4f512, please check neutron logs for more information. [ 629.423493] env[62914]: Removing descriptor: 15 [ 629.423493] env[62914]: ERROR nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e9e80771-8137-4491-b800-397ba5b4f512, please check neutron logs for more information. [ 629.423493] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Traceback (most recent call last): [ 629.423493] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 629.423493] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] yield resources [ 629.423493] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.423493] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] self.driver.spawn(context, instance, image_meta, [ 629.423493] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 629.423493] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.423493] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.423493] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] vm_ref = self.build_virtual_machine(instance, [ 629.424129] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.424129] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.424129] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.424129] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] for vif in network_info: [ 629.424129] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.424129] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] return self._sync_wrapper(fn, *args, **kwargs) [ 629.424129] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.424129] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] self.wait() [ 629.424129] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.424129] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] self[:] = self._gt.wait() [ 629.424129] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.424129] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] return self._exit_event.wait() [ 629.424129] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.424629] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] result = hub.switch() [ 629.424629] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.424629] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] return self.greenlet.switch() [ 629.424629] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.424629] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] result = function(*args, **kwargs) [ 629.424629] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 629.424629] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] return func(*args, **kwargs) [ 629.424629] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.424629] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] raise e [ 629.424629] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.424629] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] nwinfo = self.network_api.allocate_for_instance( [ 629.424629] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.424629] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] created_port_ids = self._update_ports_for_instance( [ 629.425303] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.425303] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] with excutils.save_and_reraise_exception(): [ 629.425303] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.425303] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] self.force_reraise() [ 629.425303] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.425303] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] raise self.value [ 629.425303] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.425303] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] updated_port = self._update_port( [ 629.425303] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.425303] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] _ensure_no_port_binding_failure(port) [ 629.425303] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.425303] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] raise exception.PortBindingFailed(port_id=port['id']) [ 629.425852] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] nova.exception.PortBindingFailed: Binding failed for port e9e80771-8137-4491-b800-397ba5b4f512, please check neutron logs for more information. [ 629.425852] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] [ 629.425852] env[62914]: INFO nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Terminating instance [ 629.425852] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Acquiring lock "refresh_cache-8bda7658-2b86-4869-89e2-b10d7119ae18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.779421] env[62914]: DEBUG nova.network.neutron [req-e147aa07-c0b8-4028-b2e0-56d01bf1a1b5 req-5de146f6-c23e-4296-9263-7777c996b38a service nova] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.823037] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.650s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.824055] env[62914]: DEBUG nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 629.826285] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.217s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.888534] env[62914]: DEBUG nova.network.neutron [req-e147aa07-c0b8-4028-b2e0-56d01bf1a1b5 req-5de146f6-c23e-4296-9263-7777c996b38a service nova] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.332013] env[62914]: DEBUG nova.compute.utils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 630.336212] env[62914]: DEBUG nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 630.336391] env[62914]: DEBUG nova.network.neutron [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 630.380851] env[62914]: DEBUG nova.policy [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d74ad5e0d554a6cb6f557373bf1d841', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '56111979cb10452d9d49fec61eca4694', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 630.390391] env[62914]: DEBUG oslo_concurrency.lockutils [req-e147aa07-c0b8-4028-b2e0-56d01bf1a1b5 req-5de146f6-c23e-4296-9263-7777c996b38a service nova] Releasing lock "refresh_cache-8bda7658-2b86-4869-89e2-b10d7119ae18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.390555] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Acquired lock "refresh_cache-8bda7658-2b86-4869-89e2-b10d7119ae18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.390702] env[62914]: DEBUG nova.network.neutron [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 630.712273] env[62914]: DEBUG nova.network.neutron [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Successfully created port: 5c57f21b-7d47-4d5b-96fe-2bbc4d523a5f {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 630.776156] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-599f5e38-a239-4eb7-be8e-bf78c9540801 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.783582] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98249c3d-ae7d-4751-b793-88d1b4d40166 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.812370] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d1ea35-afcd-4c9f-92ee-24ca42152517 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.819494] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0976d4fe-dbf2-4680-9ffb-0c1c3ba03946 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.832388] env[62914]: DEBUG nova.compute.provider_tree [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.839552] env[62914]: DEBUG nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 630.913288] env[62914]: DEBUG nova.network.neutron [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.003145] env[62914]: DEBUG nova.network.neutron [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.279545] env[62914]: DEBUG nova.compute.manager [req-1874c58e-9dc9-4f8d-b829-88b0d1a216c2 req-834eaa2b-2610-4496-a72c-d6a88c1d391d service nova] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Received event network-vif-deleted-e9e80771-8137-4491-b800-397ba5b4f512 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.337402] env[62914]: DEBUG nova.scheduler.client.report [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.506740] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Releasing lock "refresh_cache-8bda7658-2b86-4869-89e2-b10d7119ae18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.508222] env[62914]: DEBUG nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 631.508222] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 631.508222] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fff6aecb-ba28-4a8c-9bcb-423f8d904b54 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.521679] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748751c1-616d-47d3-bbd8-532de28274a5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.547046] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8bda7658-2b86-4869-89e2-b10d7119ae18 could not be found. [ 631.547046] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 631.547046] env[62914]: INFO nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Took 0.04 seconds to destroy the instance on the hypervisor. [ 631.547046] env[62914]: DEBUG oslo.service.loopingcall [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.547046] env[62914]: DEBUG nova.compute.manager [-] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.547046] env[62914]: DEBUG nova.network.neutron [-] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 631.559811] env[62914]: DEBUG nova.network.neutron [-] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.662288] env[62914]: ERROR nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5c57f21b-7d47-4d5b-96fe-2bbc4d523a5f, please check neutron logs for more information. [ 631.662288] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 631.662288] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.662288] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 631.662288] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.662288] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 631.662288] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.662288] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 631.662288] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.662288] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 631.662288] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.662288] env[62914]: ERROR nova.compute.manager raise self.value [ 631.662288] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.662288] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 631.662288] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.662288] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 631.662741] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.662741] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 631.662741] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5c57f21b-7d47-4d5b-96fe-2bbc4d523a5f, please check neutron logs for more information. [ 631.662741] env[62914]: ERROR nova.compute.manager [ 631.662741] env[62914]: Traceback (most recent call last): [ 631.662741] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 631.662741] env[62914]: listener.cb(fileno) [ 631.662741] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.662741] env[62914]: result = function(*args, **kwargs) [ 631.662741] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.662741] env[62914]: return func(*args, **kwargs) [ 631.662741] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.662741] env[62914]: raise e [ 631.662741] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.662741] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 631.662741] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.662741] env[62914]: created_port_ids = self._update_ports_for_instance( [ 631.662741] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.662741] env[62914]: with excutils.save_and_reraise_exception(): [ 631.662741] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.662741] env[62914]: self.force_reraise() [ 631.662741] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.662741] env[62914]: raise self.value [ 631.662741] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.662741] env[62914]: updated_port = self._update_port( [ 631.662741] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.662741] env[62914]: _ensure_no_port_binding_failure(port) [ 631.662741] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.662741] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 631.663453] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 5c57f21b-7d47-4d5b-96fe-2bbc4d523a5f, please check neutron logs for more information. [ 631.663453] env[62914]: Removing descriptor: 15 [ 631.846030] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.018s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.846030] env[62914]: ERROR nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1194b118-455c-4a76-a106-cde6931bec18, please check neutron logs for more information. [ 631.846030] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Traceback (most recent call last): [ 631.846030] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.846030] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] self.driver.spawn(context, instance, image_meta, [ 631.846030] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 631.846030] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.846030] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.846030] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] vm_ref = self.build_virtual_machine(instance, [ 631.846352] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.846352] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.846352] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.846352] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] for vif in network_info: [ 631.846352] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.846352] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] return self._sync_wrapper(fn, *args, **kwargs) [ 631.846352] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.846352] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] self.wait() [ 631.846352] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.846352] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] self[:] = self._gt.wait() [ 631.846352] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.846352] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] return self._exit_event.wait() [ 631.846352] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.846654] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] result = hub.switch() [ 631.846654] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.846654] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] return self.greenlet.switch() [ 631.846654] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.846654] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] result = function(*args, **kwargs) [ 631.846654] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.846654] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] return func(*args, **kwargs) [ 631.846654] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.846654] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] raise e [ 631.846654] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.846654] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] nwinfo = self.network_api.allocate_for_instance( [ 631.846654] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.846654] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] created_port_ids = self._update_ports_for_instance( [ 631.846962] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.846962] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] with excutils.save_and_reraise_exception(): [ 631.846962] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.846962] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] self.force_reraise() [ 631.846962] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.846962] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] raise self.value [ 631.846962] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.846962] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] updated_port = self._update_port( [ 631.846962] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.846962] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] _ensure_no_port_binding_failure(port) [ 631.846962] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.846962] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] raise exception.PortBindingFailed(port_id=port['id']) [ 631.847251] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] nova.exception.PortBindingFailed: Binding failed for port 1194b118-455c-4a76-a106-cde6931bec18, please check neutron logs for more information. [ 631.847251] env[62914]: ERROR nova.compute.manager [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] [ 631.847251] env[62914]: DEBUG nova.compute.utils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Binding failed for port 1194b118-455c-4a76-a106-cde6931bec18, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 631.847336] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.679s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.850369] env[62914]: DEBUG nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Build of instance 8213bc31-1bd5-40b6-99a0-1254b9517b6a was re-scheduled: Binding failed for port 1194b118-455c-4a76-a106-cde6931bec18, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 631.850610] env[62914]: DEBUG nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 631.850831] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Acquiring lock "refresh_cache-8213bc31-1bd5-40b6-99a0-1254b9517b6a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.850975] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Acquired lock "refresh_cache-8213bc31-1bd5-40b6-99a0-1254b9517b6a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.851140] env[62914]: DEBUG nova.network.neutron [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.853037] env[62914]: DEBUG nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 631.877614] env[62914]: DEBUG nova.virt.hardware [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 631.877852] env[62914]: DEBUG nova.virt.hardware [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 631.878015] env[62914]: DEBUG nova.virt.hardware [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 631.878206] env[62914]: DEBUG nova.virt.hardware [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 631.878346] env[62914]: DEBUG nova.virt.hardware [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 631.878488] env[62914]: DEBUG nova.virt.hardware [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 631.878691] env[62914]: DEBUG nova.virt.hardware [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 631.878843] env[62914]: DEBUG nova.virt.hardware [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 631.879019] env[62914]: DEBUG nova.virt.hardware [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 631.879188] env[62914]: DEBUG nova.virt.hardware [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 631.879379] env[62914]: DEBUG nova.virt.hardware [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 631.880195] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70602a8-a456-4e73-82ae-0afe66a5c8fb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.889262] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530425ca-789a-4ac4-9d65-b3ec6f68d7b8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.902699] env[62914]: ERROR nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5c57f21b-7d47-4d5b-96fe-2bbc4d523a5f, please check neutron logs for more information. [ 631.902699] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Traceback (most recent call last): [ 631.902699] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 631.902699] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] yield resources [ 631.902699] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.902699] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] self.driver.spawn(context, instance, image_meta, [ 631.902699] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 631.902699] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.902699] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.902699] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] vm_ref = self.build_virtual_machine(instance, [ 631.902699] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.903048] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.903048] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.903048] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] for vif in network_info: [ 631.903048] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.903048] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] return self._sync_wrapper(fn, *args, **kwargs) [ 631.903048] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.903048] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] self.wait() [ 631.903048] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.903048] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] self[:] = self._gt.wait() [ 631.903048] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.903048] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] return self._exit_event.wait() [ 631.903048] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 631.903048] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] current.throw(*self._exc) [ 631.903509] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.903509] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] result = function(*args, **kwargs) [ 631.903509] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.903509] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] return func(*args, **kwargs) [ 631.903509] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.903509] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] raise e [ 631.903509] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.903509] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] nwinfo = self.network_api.allocate_for_instance( [ 631.903509] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.903509] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] created_port_ids = self._update_ports_for_instance( [ 631.903509] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.903509] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] with excutils.save_and_reraise_exception(): [ 631.903509] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.903885] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] self.force_reraise() [ 631.903885] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.903885] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] raise self.value [ 631.903885] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.903885] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] updated_port = self._update_port( [ 631.903885] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.903885] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] _ensure_no_port_binding_failure(port) [ 631.903885] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.903885] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] raise exception.PortBindingFailed(port_id=port['id']) [ 631.903885] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] nova.exception.PortBindingFailed: Binding failed for port 5c57f21b-7d47-4d5b-96fe-2bbc4d523a5f, please check neutron logs for more information. [ 631.903885] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] [ 631.903885] env[62914]: INFO nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Terminating instance [ 631.904848] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Acquiring lock "refresh_cache-15938dfe-a697-450b-8c88-ff4ef703ee1f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.905008] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Acquired lock "refresh_cache-15938dfe-a697-450b-8c88-ff4ef703ee1f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.905899] env[62914]: DEBUG nova.network.neutron [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 632.062980] env[62914]: DEBUG nova.network.neutron [-] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.370790] env[62914]: DEBUG nova.network.neutron [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.424774] env[62914]: DEBUG nova.network.neutron [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.427026] env[62914]: DEBUG nova.network.neutron [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.512043] env[62914]: DEBUG nova.network.neutron [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.565243] env[62914]: INFO nova.compute.manager [-] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Took 1.02 seconds to deallocate network for instance. [ 632.570792] env[62914]: DEBUG nova.compute.claims [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 632.571018] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.740871] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d02ed73-8cff-4617-9d6f-aec44a81bd25 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.749176] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad05480-4754-414d-bf65-a9c465b7d44d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.783400] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7befb90-bf74-4870-8449-b49773a8e958 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.791069] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b1c217-2327-4894-9cf0-0eb214b1e25d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.804598] env[62914]: DEBUG nova.compute.provider_tree [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.929789] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Releasing lock "refresh_cache-8213bc31-1bd5-40b6-99a0-1254b9517b6a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.930048] env[62914]: DEBUG nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 632.930218] env[62914]: DEBUG nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.930383] env[62914]: DEBUG nova.network.neutron [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.946025] env[62914]: DEBUG nova.network.neutron [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.015278] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Releasing lock "refresh_cache-15938dfe-a697-450b-8c88-ff4ef703ee1f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.015415] env[62914]: DEBUG nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 633.015679] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 633.015965] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-87d36515-e0f1-4eab-9a96-f0128f254d63 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.024503] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52bb0f8e-b289-4c8f-a3b3-dc725f64bd7c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.047177] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 15938dfe-a697-450b-8c88-ff4ef703ee1f could not be found. [ 633.047404] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 633.047586] env[62914]: INFO nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 633.047829] env[62914]: DEBUG oslo.service.loopingcall [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.048105] env[62914]: DEBUG nova.compute.manager [-] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.048165] env[62914]: DEBUG nova.network.neutron [-] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 633.071932] env[62914]: DEBUG nova.network.neutron [-] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.307585] env[62914]: DEBUG nova.scheduler.client.report [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.342511] env[62914]: DEBUG nova.compute.manager [req-950ba01d-ba66-4391-92d8-a41cf1dfa94d req-038ce301-0e02-4beb-bc1d-92cf106b2ac7 service nova] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Received event network-changed-5c57f21b-7d47-4d5b-96fe-2bbc4d523a5f {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 633.342721] env[62914]: DEBUG nova.compute.manager [req-950ba01d-ba66-4391-92d8-a41cf1dfa94d req-038ce301-0e02-4beb-bc1d-92cf106b2ac7 service nova] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Refreshing instance network info cache due to event network-changed-5c57f21b-7d47-4d5b-96fe-2bbc4d523a5f. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 633.343050] env[62914]: DEBUG oslo_concurrency.lockutils [req-950ba01d-ba66-4391-92d8-a41cf1dfa94d req-038ce301-0e02-4beb-bc1d-92cf106b2ac7 service nova] Acquiring lock "refresh_cache-15938dfe-a697-450b-8c88-ff4ef703ee1f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.343107] env[62914]: DEBUG oslo_concurrency.lockutils [req-950ba01d-ba66-4391-92d8-a41cf1dfa94d req-038ce301-0e02-4beb-bc1d-92cf106b2ac7 service nova] Acquired lock "refresh_cache-15938dfe-a697-450b-8c88-ff4ef703ee1f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.343264] env[62914]: DEBUG nova.network.neutron [req-950ba01d-ba66-4391-92d8-a41cf1dfa94d req-038ce301-0e02-4beb-bc1d-92cf106b2ac7 service nova] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Refreshing network info cache for port 5c57f21b-7d47-4d5b-96fe-2bbc4d523a5f {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 633.447729] env[62914]: DEBUG nova.network.neutron [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.574592] env[62914]: DEBUG nova.network.neutron [-] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.812787] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.965s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.813470] env[62914]: ERROR nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 691b9a22-1a0a-4a84-a136-b828743aff9c, please check neutron logs for more information. [ 633.813470] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Traceback (most recent call last): [ 633.813470] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 633.813470] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] self.driver.spawn(context, instance, image_meta, [ 633.813470] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 633.813470] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.813470] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.813470] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] vm_ref = self.build_virtual_machine(instance, [ 633.813470] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.813470] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.813470] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.813829] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] for vif in network_info: [ 633.813829] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.813829] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] return self._sync_wrapper(fn, *args, **kwargs) [ 633.813829] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.813829] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] self.wait() [ 633.813829] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.813829] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] self[:] = self._gt.wait() [ 633.813829] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.813829] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] return self._exit_event.wait() [ 633.813829] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.813829] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] result = hub.switch() [ 633.813829] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.813829] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] return self.greenlet.switch() [ 633.814210] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.814210] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] result = function(*args, **kwargs) [ 633.814210] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 633.814210] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] return func(*args, **kwargs) [ 633.814210] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.814210] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] raise e [ 633.814210] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.814210] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] nwinfo = self.network_api.allocate_for_instance( [ 633.814210] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.814210] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] created_port_ids = self._update_ports_for_instance( [ 633.814210] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.814210] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] with excutils.save_and_reraise_exception(): [ 633.814210] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.814626] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] self.force_reraise() [ 633.814626] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.814626] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] raise self.value [ 633.814626] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.814626] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] updated_port = self._update_port( [ 633.814626] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.814626] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] _ensure_no_port_binding_failure(port) [ 633.814626] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.814626] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] raise exception.PortBindingFailed(port_id=port['id']) [ 633.814626] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] nova.exception.PortBindingFailed: Binding failed for port 691b9a22-1a0a-4a84-a136-b828743aff9c, please check neutron logs for more information. [ 633.814626] env[62914]: ERROR nova.compute.manager [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] [ 633.814936] env[62914]: DEBUG nova.compute.utils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Binding failed for port 691b9a22-1a0a-4a84-a136-b828743aff9c, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 633.815607] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.362s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.815831] env[62914]: DEBUG nova.objects.instance [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Lazy-loading 'resources' on Instance uuid e11079b9-5f15-41d6-8c8a-c08dc7f63b2f {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 633.817227] env[62914]: DEBUG nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Build of instance 5e695914-afa5-4f79-972d-852a4ac8aa20 was re-scheduled: Binding failed for port 691b9a22-1a0a-4a84-a136-b828743aff9c, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 633.817503] env[62914]: DEBUG nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 633.817725] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Acquiring lock "refresh_cache-5e695914-afa5-4f79-972d-852a4ac8aa20" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.817874] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Acquired lock "refresh_cache-5e695914-afa5-4f79-972d-852a4ac8aa20" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.818044] env[62914]: DEBUG nova.network.neutron [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.862651] env[62914]: DEBUG nova.network.neutron [req-950ba01d-ba66-4391-92d8-a41cf1dfa94d req-038ce301-0e02-4beb-bc1d-92cf106b2ac7 service nova] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.944585] env[62914]: DEBUG nova.network.neutron [req-950ba01d-ba66-4391-92d8-a41cf1dfa94d req-038ce301-0e02-4beb-bc1d-92cf106b2ac7 service nova] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.950602] env[62914]: INFO nova.compute.manager [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] [instance: 8213bc31-1bd5-40b6-99a0-1254b9517b6a] Took 1.02 seconds to deallocate network for instance. [ 634.077026] env[62914]: INFO nova.compute.manager [-] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Took 1.03 seconds to deallocate network for instance. [ 634.078919] env[62914]: DEBUG nova.compute.claims [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 634.079112] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.335715] env[62914]: DEBUG nova.network.neutron [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.421228] env[62914]: DEBUG nova.network.neutron [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.447402] env[62914]: DEBUG oslo_concurrency.lockutils [req-950ba01d-ba66-4391-92d8-a41cf1dfa94d req-038ce301-0e02-4beb-bc1d-92cf106b2ac7 service nova] Releasing lock "refresh_cache-15938dfe-a697-450b-8c88-ff4ef703ee1f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.447643] env[62914]: DEBUG nova.compute.manager [req-950ba01d-ba66-4391-92d8-a41cf1dfa94d req-038ce301-0e02-4beb-bc1d-92cf106b2ac7 service nova] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Received event network-vif-deleted-5c57f21b-7d47-4d5b-96fe-2bbc4d523a5f {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.706507] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50bff884-77eb-49bc-8d83-7358b0b643fc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.714737] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c43bc3-145a-4e9e-8129-50a78d4c087b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.743841] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b207978-7586-488c-a29f-c638d814ef31 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.750831] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb9b77b-a37a-4fc1-920d-144256c0e019 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.764721] env[62914]: DEBUG nova.compute.provider_tree [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.926908] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Releasing lock "refresh_cache-5e695914-afa5-4f79-972d-852a4ac8aa20" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.926908] env[62914]: DEBUG nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 634.926908] env[62914]: DEBUG nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.926908] env[62914]: DEBUG nova.network.neutron [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 634.939121] env[62914]: DEBUG nova.network.neutron [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.979526] env[62914]: INFO nova.scheduler.client.report [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Deleted allocations for instance 8213bc31-1bd5-40b6-99a0-1254b9517b6a [ 635.268383] env[62914]: DEBUG nova.scheduler.client.report [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 635.441612] env[62914]: DEBUG nova.network.neutron [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.489483] env[62914]: DEBUG oslo_concurrency.lockutils [None req-96ea77b5-b0e1-43f0-8466-866f34f543cd tempest-ServerDiagnosticsTest-422723390 tempest-ServerDiagnosticsTest-422723390-project-member] Lock "8213bc31-1bd5-40b6-99a0-1254b9517b6a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.847s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.773626] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.958s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.777161] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.021s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.777965] env[62914]: INFO nova.compute.claims [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 635.803703] env[62914]: INFO nova.scheduler.client.report [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Deleted allocations for instance e11079b9-5f15-41d6-8c8a-c08dc7f63b2f [ 635.945368] env[62914]: INFO nova.compute.manager [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] [instance: 5e695914-afa5-4f79-972d-852a4ac8aa20] Took 1.02 seconds to deallocate network for instance. [ 635.991761] env[62914]: DEBUG nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 636.312489] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a93ba577-9f28-4ed1-b630-551e991fde64 tempest-ServersAaction247Test-1885368857 tempest-ServersAaction247Test-1885368857-project-member] Lock "e11079b9-5f15-41d6-8c8a-c08dc7f63b2f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.564s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.516244] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.976211] env[62914]: INFO nova.scheduler.client.report [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Deleted allocations for instance 5e695914-afa5-4f79-972d-852a4ac8aa20 [ 637.219960] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79907c89-8a5e-4f47-9743-bc8d40a65bc9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.231514] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57115c00-078a-4ea6-94df-f30d58646b45 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.261056] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a65b3d6d-de38-4fe2-82ba-494599598811 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.270260] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a972d598-d5d3-4b94-9579-89513a75ee50 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.283956] env[62914]: DEBUG nova.compute.provider_tree [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.490598] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bab92ea3-4306-43c2-b169-bcea29457655 tempest-InstanceActionsTestJSON-1579593933 tempest-InstanceActionsTestJSON-1579593933-project-member] Lock "5e695914-afa5-4f79-972d-852a4ac8aa20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.216s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.792101] env[62914]: DEBUG nova.scheduler.client.report [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.999129] env[62914]: DEBUG nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 638.297641] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.298518] env[62914]: DEBUG nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 638.301899] env[62914]: DEBUG oslo_concurrency.lockutils [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.112s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.302358] env[62914]: DEBUG nova.objects.instance [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62914) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 638.562207] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.807173] env[62914]: DEBUG nova.compute.utils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 638.812975] env[62914]: DEBUG nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 638.813490] env[62914]: DEBUG nova.network.neutron [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 638.911417] env[62914]: DEBUG nova.policy [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '456f87454d3c484ea80e8ddf6b565cfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8e87bdcbeea4e9ea2b08630acc32ab1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 639.318355] env[62914]: DEBUG oslo_concurrency.lockutils [None req-21c2018e-432c-459e-b591-0814f201e010 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.319662] env[62914]: DEBUG nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 639.322202] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.136s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.776215] env[62914]: DEBUG nova.network.neutron [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Successfully created port: 3589f50e-46e4-4a1f-8b9d-51955b48102d {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 640.337529] env[62914]: DEBUG nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 640.406338] env[62914]: DEBUG nova.virt.hardware [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 640.406338] env[62914]: DEBUG nova.virt.hardware [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 640.407057] env[62914]: DEBUG nova.virt.hardware [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 640.407444] env[62914]: DEBUG nova.virt.hardware [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 640.408203] env[62914]: DEBUG nova.virt.hardware [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 640.408852] env[62914]: DEBUG nova.virt.hardware [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 640.408852] env[62914]: DEBUG nova.virt.hardware [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 640.408852] env[62914]: DEBUG nova.virt.hardware [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 640.408951] env[62914]: DEBUG nova.virt.hardware [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 640.409157] env[62914]: DEBUG nova.virt.hardware [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 640.410085] env[62914]: DEBUG nova.virt.hardware [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 640.410348] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2c6ca4-a560-49f0-899d-f71869377410 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.423254] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcf2105-b256-494c-93ab-04a89a66c879 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.529419] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56970fae-08ae-4e0d-9683-46bece9c6520 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.537363] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fdb4d7-9db5-457d-b7c2-acee2401e203 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.576830] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcfb79b-b5ad-4752-9282-e66157522e2a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.586066] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb95724-af9d-4101-a3cc-264d8eefafaf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.601344] env[62914]: DEBUG nova.compute.provider_tree [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.106656] env[62914]: DEBUG nova.scheduler.client.report [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.385471] env[62914]: DEBUG nova.compute.manager [req-19cb7c57-5f19-4c9b-9334-301a8f8aa48a req-17064d26-d61c-43af-819b-235b94b3abe3 service nova] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Received event network-changed-3589f50e-46e4-4a1f-8b9d-51955b48102d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 641.385787] env[62914]: DEBUG nova.compute.manager [req-19cb7c57-5f19-4c9b-9334-301a8f8aa48a req-17064d26-d61c-43af-819b-235b94b3abe3 service nova] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Refreshing instance network info cache due to event network-changed-3589f50e-46e4-4a1f-8b9d-51955b48102d. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 641.385903] env[62914]: DEBUG oslo_concurrency.lockutils [req-19cb7c57-5f19-4c9b-9334-301a8f8aa48a req-17064d26-d61c-43af-819b-235b94b3abe3 service nova] Acquiring lock "refresh_cache-91328e42-ba3c-4d63-a805-973c2bfa2afb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.387294] env[62914]: DEBUG oslo_concurrency.lockutils [req-19cb7c57-5f19-4c9b-9334-301a8f8aa48a req-17064d26-d61c-43af-819b-235b94b3abe3 service nova] Acquired lock "refresh_cache-91328e42-ba3c-4d63-a805-973c2bfa2afb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.387294] env[62914]: DEBUG nova.network.neutron [req-19cb7c57-5f19-4c9b-9334-301a8f8aa48a req-17064d26-d61c-43af-819b-235b94b3abe3 service nova] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Refreshing network info cache for port 3589f50e-46e4-4a1f-8b9d-51955b48102d {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 641.611911] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.290s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.615899] env[62914]: ERROR nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c886397b-a988-45ab-9cdc-21a3afc92a24, please check neutron logs for more information. [ 641.615899] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] Traceback (most recent call last): [ 641.615899] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.615899] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] self.driver.spawn(context, instance, image_meta, [ 641.615899] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 641.615899] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.615899] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.615899] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] vm_ref = self.build_virtual_machine(instance, [ 641.615899] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.615899] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.615899] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.616250] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] for vif in network_info: [ 641.616250] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.616250] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] return self._sync_wrapper(fn, *args, **kwargs) [ 641.616250] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.616250] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] self.wait() [ 641.616250] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.616250] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] self[:] = self._gt.wait() [ 641.616250] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.616250] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] return self._exit_event.wait() [ 641.616250] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 641.616250] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] result = hub.switch() [ 641.616250] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 641.616250] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] return self.greenlet.switch() [ 641.616552] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.616552] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] result = function(*args, **kwargs) [ 641.616552] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 641.616552] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] return func(*args, **kwargs) [ 641.616552] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.616552] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] raise e [ 641.616552] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.616552] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] nwinfo = self.network_api.allocate_for_instance( [ 641.616552] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.616552] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] created_port_ids = self._update_ports_for_instance( [ 641.616552] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.616552] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] with excutils.save_and_reraise_exception(): [ 641.616552] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.616853] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] self.force_reraise() [ 641.616853] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.616853] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] raise self.value [ 641.616853] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.616853] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] updated_port = self._update_port( [ 641.616853] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.616853] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] _ensure_no_port_binding_failure(port) [ 641.616853] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.616853] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] raise exception.PortBindingFailed(port_id=port['id']) [ 641.616853] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] nova.exception.PortBindingFailed: Binding failed for port c886397b-a988-45ab-9cdc-21a3afc92a24, please check neutron logs for more information. [ 641.616853] env[62914]: ERROR nova.compute.manager [instance: 964f98e4-3462-4362-b551-576302776f3a] [ 641.617130] env[62914]: DEBUG nova.compute.utils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Binding failed for port c886397b-a988-45ab-9cdc-21a3afc92a24, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 641.619074] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.137s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.620731] env[62914]: INFO nova.compute.claims [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 641.623443] env[62914]: DEBUG nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Build of instance 964f98e4-3462-4362-b551-576302776f3a was re-scheduled: Binding failed for port c886397b-a988-45ab-9cdc-21a3afc92a24, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 641.623930] env[62914]: DEBUG nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 641.624171] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Acquiring lock "refresh_cache-964f98e4-3462-4362-b551-576302776f3a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.624317] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Acquired lock "refresh_cache-964f98e4-3462-4362-b551-576302776f3a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.624468] env[62914]: DEBUG nova.network.neutron [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.859293] env[62914]: ERROR nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3589f50e-46e4-4a1f-8b9d-51955b48102d, please check neutron logs for more information. [ 641.859293] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 641.859293] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.859293] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 641.859293] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.859293] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 641.859293] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.859293] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 641.859293] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.859293] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 641.859293] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.859293] env[62914]: ERROR nova.compute.manager raise self.value [ 641.859293] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.859293] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 641.859293] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.859293] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 641.859705] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.859705] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 641.859705] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3589f50e-46e4-4a1f-8b9d-51955b48102d, please check neutron logs for more information. [ 641.859705] env[62914]: ERROR nova.compute.manager [ 641.859705] env[62914]: Traceback (most recent call last): [ 641.859705] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 641.859705] env[62914]: listener.cb(fileno) [ 641.859705] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.859705] env[62914]: result = function(*args, **kwargs) [ 641.859705] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 641.859705] env[62914]: return func(*args, **kwargs) [ 641.859705] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.859705] env[62914]: raise e [ 641.859705] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.859705] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 641.859705] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.859705] env[62914]: created_port_ids = self._update_ports_for_instance( [ 641.859705] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.859705] env[62914]: with excutils.save_and_reraise_exception(): [ 641.859705] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.859705] env[62914]: self.force_reraise() [ 641.859705] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.859705] env[62914]: raise self.value [ 641.859705] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.859705] env[62914]: updated_port = self._update_port( [ 641.859705] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.859705] env[62914]: _ensure_no_port_binding_failure(port) [ 641.859705] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.859705] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 641.860597] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 3589f50e-46e4-4a1f-8b9d-51955b48102d, please check neutron logs for more information. [ 641.860597] env[62914]: Removing descriptor: 19 [ 641.860597] env[62914]: ERROR nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3589f50e-46e4-4a1f-8b9d-51955b48102d, please check neutron logs for more information. [ 641.860597] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Traceback (most recent call last): [ 641.860597] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 641.860597] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] yield resources [ 641.860597] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.860597] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] self.driver.spawn(context, instance, image_meta, [ 641.860597] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 641.860597] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.860597] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.860597] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] vm_ref = self.build_virtual_machine(instance, [ 641.860949] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.860949] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.860949] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.860949] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] for vif in network_info: [ 641.860949] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.860949] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] return self._sync_wrapper(fn, *args, **kwargs) [ 641.860949] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.860949] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] self.wait() [ 641.860949] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.860949] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] self[:] = self._gt.wait() [ 641.860949] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.860949] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] return self._exit_event.wait() [ 641.860949] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 641.861264] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] result = hub.switch() [ 641.861264] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 641.861264] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] return self.greenlet.switch() [ 641.861264] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.861264] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] result = function(*args, **kwargs) [ 641.861264] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 641.861264] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] return func(*args, **kwargs) [ 641.861264] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.861264] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] raise e [ 641.861264] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.861264] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] nwinfo = self.network_api.allocate_for_instance( [ 641.861264] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.861264] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] created_port_ids = self._update_ports_for_instance( [ 641.861564] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.861564] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] with excutils.save_and_reraise_exception(): [ 641.861564] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.861564] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] self.force_reraise() [ 641.861564] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.861564] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] raise self.value [ 641.861564] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.861564] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] updated_port = self._update_port( [ 641.861564] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.861564] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] _ensure_no_port_binding_failure(port) [ 641.861564] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.861564] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] raise exception.PortBindingFailed(port_id=port['id']) [ 641.861882] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] nova.exception.PortBindingFailed: Binding failed for port 3589f50e-46e4-4a1f-8b9d-51955b48102d, please check neutron logs for more information. [ 641.861882] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] [ 641.861882] env[62914]: INFO nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Terminating instance [ 641.866707] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Acquiring lock "refresh_cache-91328e42-ba3c-4d63-a805-973c2bfa2afb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.916111] env[62914]: DEBUG nova.network.neutron [req-19cb7c57-5f19-4c9b-9334-301a8f8aa48a req-17064d26-d61c-43af-819b-235b94b3abe3 service nova] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.991023] env[62914]: DEBUG nova.network.neutron [req-19cb7c57-5f19-4c9b-9334-301a8f8aa48a req-17064d26-d61c-43af-819b-235b94b3abe3 service nova] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.170031] env[62914]: DEBUG nova.network.neutron [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.288588] env[62914]: DEBUG nova.network.neutron [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.495112] env[62914]: DEBUG oslo_concurrency.lockutils [req-19cb7c57-5f19-4c9b-9334-301a8f8aa48a req-17064d26-d61c-43af-819b-235b94b3abe3 service nova] Releasing lock "refresh_cache-91328e42-ba3c-4d63-a805-973c2bfa2afb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.495112] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Acquired lock "refresh_cache-91328e42-ba3c-4d63-a805-973c2bfa2afb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.495112] env[62914]: DEBUG nova.network.neutron [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 642.760642] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "39489032-5430-471c-8d44-2c74682ed03b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.760952] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "39489032-5430-471c-8d44-2c74682ed03b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.776463] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Acquiring lock "6201e8b7-ca3d-4eb0-8c45-67f54e66173c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.776707] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Lock "6201e8b7-ca3d-4eb0-8c45-67f54e66173c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.790489] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Releasing lock "refresh_cache-964f98e4-3462-4362-b551-576302776f3a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.790704] env[62914]: DEBUG nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 642.790874] env[62914]: DEBUG nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.791268] env[62914]: DEBUG nova.network.neutron [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 642.810634] env[62914]: DEBUG nova.network.neutron [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.014451] env[62914]: DEBUG nova.network.neutron [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.070935] env[62914]: DEBUG nova.network.neutron [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.106768] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d4c3ae-9be1-4fa8-8f7a-55c1893cc056 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.114666] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a040ac-c5c3-40b0-a650-1097c5bc7fc4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.146303] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59879e1a-7caf-4ed6-9777-a8a9c4456a72 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.153787] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c23da0c-5fee-4aea-aab0-ee7754c0e2f8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.167347] env[62914]: DEBUG nova.compute.provider_tree [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.317207] env[62914]: DEBUG nova.network.neutron [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.435702] env[62914]: DEBUG nova.compute.manager [req-63b9b879-d876-42a6-a77e-ceed39574865 req-f08e24bb-949e-4bde-a751-6b3afaecc824 service nova] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Received event network-vif-deleted-3589f50e-46e4-4a1f-8b9d-51955b48102d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 643.573696] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Releasing lock "refresh_cache-91328e42-ba3c-4d63-a805-973c2bfa2afb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.574324] env[62914]: DEBUG nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 643.574499] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 643.574751] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0bb1a4d1-ed9a-49e1-8e01-412e5d183e8a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.584306] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5b079c-eb56-4727-9d7e-2c4786546fa8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.610456] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 91328e42-ba3c-4d63-a805-973c2bfa2afb could not be found. [ 643.610679] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 643.610858] env[62914]: INFO nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 643.611514] env[62914]: DEBUG oslo.service.loopingcall [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.611514] env[62914]: DEBUG nova.compute.manager [-] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.611514] env[62914]: DEBUG nova.network.neutron [-] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.636953] env[62914]: DEBUG nova.network.neutron [-] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.670621] env[62914]: DEBUG nova.scheduler.client.report [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 643.819350] env[62914]: INFO nova.compute.manager [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] [instance: 964f98e4-3462-4362-b551-576302776f3a] Took 1.03 seconds to deallocate network for instance. [ 644.139391] env[62914]: DEBUG nova.network.neutron [-] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.175092] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.556s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.175579] env[62914]: DEBUG nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 644.178315] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.653s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.178526] env[62914]: DEBUG nova.objects.instance [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lazy-loading 'resources' on Instance uuid f3509973-013e-45ce-87f9-357e782f26d7 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 644.642592] env[62914]: INFO nova.compute.manager [-] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Took 1.03 seconds to deallocate network for instance. [ 644.647508] env[62914]: DEBUG nova.compute.claims [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 644.647711] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.684822] env[62914]: DEBUG nova.compute.utils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 644.686279] env[62914]: DEBUG nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 644.686452] env[62914]: DEBUG nova.network.neutron [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 644.744956] env[62914]: DEBUG nova.policy [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb1db37559a64d41961aa212b8e8b282', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53fb59403531443fb1452ff0e7b699c6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 644.864862] env[62914]: INFO nova.scheduler.client.report [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Deleted allocations for instance 964f98e4-3462-4362-b551-576302776f3a [ 645.058707] env[62914]: DEBUG nova.network.neutron [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Successfully created port: 8d185603-cceb-42c2-870a-7c89e54dff99 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 645.070530] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2878bc8b-0965-45a2-b97a-06f6d113e8e3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.078055] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4c3f91-1919-4605-95ea-912e102383da {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.111788] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cbde633-a139-4b93-a341-6935f0dd25a0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.119375] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c963c118-b296-4da7-98c2-79bbd02ff0d2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.132177] env[62914]: DEBUG nova.compute.provider_tree [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.189250] env[62914]: DEBUG nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 645.374971] env[62914]: DEBUG oslo_concurrency.lockutils [None req-00b7f41c-bf2d-4a26-b7ca-bacf01e18e99 tempest-InstanceActionsNegativeTestJSON-1634370216 tempest-InstanceActionsNegativeTestJSON-1634370216-project-member] Lock "964f98e4-3462-4362-b551-576302776f3a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.210s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.635422] env[62914]: DEBUG nova.scheduler.client.report [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.877526] env[62914]: DEBUG nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 645.977280] env[62914]: DEBUG nova.compute.manager [req-16f367d5-2a5c-46c5-bce6-02144e4f4d0e req-c80e6a2c-e6e1-4989-a576-e876b4cd78b8 service nova] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Received event network-changed-8d185603-cceb-42c2-870a-7c89e54dff99 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 645.977403] env[62914]: DEBUG nova.compute.manager [req-16f367d5-2a5c-46c5-bce6-02144e4f4d0e req-c80e6a2c-e6e1-4989-a576-e876b4cd78b8 service nova] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Refreshing instance network info cache due to event network-changed-8d185603-cceb-42c2-870a-7c89e54dff99. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 645.977636] env[62914]: DEBUG oslo_concurrency.lockutils [req-16f367d5-2a5c-46c5-bce6-02144e4f4d0e req-c80e6a2c-e6e1-4989-a576-e876b4cd78b8 service nova] Acquiring lock "refresh_cache-0419073e-5340-4805-a1e0-417ccf5fdd30" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.977751] env[62914]: DEBUG oslo_concurrency.lockutils [req-16f367d5-2a5c-46c5-bce6-02144e4f4d0e req-c80e6a2c-e6e1-4989-a576-e876b4cd78b8 service nova] Acquired lock "refresh_cache-0419073e-5340-4805-a1e0-417ccf5fdd30" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.977902] env[62914]: DEBUG nova.network.neutron [req-16f367d5-2a5c-46c5-bce6-02144e4f4d0e req-c80e6a2c-e6e1-4989-a576-e876b4cd78b8 service nova] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Refreshing network info cache for port 8d185603-cceb-42c2-870a-7c89e54dff99 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 646.105703] env[62914]: ERROR nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8d185603-cceb-42c2-870a-7c89e54dff99, please check neutron logs for more information. [ 646.105703] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 646.105703] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.105703] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 646.105703] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.105703] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 646.105703] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.105703] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 646.105703] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.105703] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 646.105703] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.105703] env[62914]: ERROR nova.compute.manager raise self.value [ 646.105703] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.105703] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 646.105703] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.105703] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 646.106487] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.106487] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 646.106487] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8d185603-cceb-42c2-870a-7c89e54dff99, please check neutron logs for more information. [ 646.106487] env[62914]: ERROR nova.compute.manager [ 646.106487] env[62914]: Traceback (most recent call last): [ 646.106487] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 646.106487] env[62914]: listener.cb(fileno) [ 646.106487] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.106487] env[62914]: result = function(*args, **kwargs) [ 646.106487] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 646.106487] env[62914]: return func(*args, **kwargs) [ 646.106487] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.106487] env[62914]: raise e [ 646.106487] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.106487] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 646.106487] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.106487] env[62914]: created_port_ids = self._update_ports_for_instance( [ 646.106487] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.106487] env[62914]: with excutils.save_and_reraise_exception(): [ 646.106487] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.106487] env[62914]: self.force_reraise() [ 646.106487] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.106487] env[62914]: raise self.value [ 646.106487] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.106487] env[62914]: updated_port = self._update_port( [ 646.106487] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.106487] env[62914]: _ensure_no_port_binding_failure(port) [ 646.106487] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.106487] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 646.107502] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 8d185603-cceb-42c2-870a-7c89e54dff99, please check neutron logs for more information. [ 646.107502] env[62914]: Removing descriptor: 19 [ 646.145929] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.967s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.148687] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 21.994s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.148788] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.148864] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62914) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 646.149258] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.742s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.152505] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d47d65e9-d21c-40bd-9bbd-3ac67b65e17a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.160708] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92892808-0245-4196-84cf-d83198e49d67 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.167029] env[62914]: INFO nova.scheduler.client.report [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Deleted allocations for instance f3509973-013e-45ce-87f9-357e782f26d7 [ 646.177907] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b04291-e786-4899-8419-ce1384bc260c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.184948] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd12666-8ceb-42a3-ab71-2880b478c236 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.216870] env[62914]: DEBUG nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 646.218882] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181236MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62914) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 646.219048] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.242060] env[62914]: DEBUG nova.virt.hardware [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 646.242304] env[62914]: DEBUG nova.virt.hardware [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 646.242457] env[62914]: DEBUG nova.virt.hardware [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 646.242633] env[62914]: DEBUG nova.virt.hardware [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 646.242777] env[62914]: DEBUG nova.virt.hardware [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 646.242922] env[62914]: DEBUG nova.virt.hardware [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 646.243328] env[62914]: DEBUG nova.virt.hardware [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 646.243521] env[62914]: DEBUG nova.virt.hardware [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 646.243717] env[62914]: DEBUG nova.virt.hardware [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 646.243896] env[62914]: DEBUG nova.virt.hardware [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 646.244086] env[62914]: DEBUG nova.virt.hardware [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 646.244871] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2ac294-5423-4d9d-9df4-4985cfb3fa9c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.252513] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512ce8ed-a957-4d0b-a3a7-8ef34fbd6a05 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.267372] env[62914]: ERROR nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8d185603-cceb-42c2-870a-7c89e54dff99, please check neutron logs for more information. [ 646.267372] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Traceback (most recent call last): [ 646.267372] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 646.267372] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] yield resources [ 646.267372] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 646.267372] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] self.driver.spawn(context, instance, image_meta, [ 646.267372] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 646.267372] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.267372] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.267372] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] vm_ref = self.build_virtual_machine(instance, [ 646.267372] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.267760] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.267760] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.267760] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] for vif in network_info: [ 646.267760] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 646.267760] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] return self._sync_wrapper(fn, *args, **kwargs) [ 646.267760] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 646.267760] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] self.wait() [ 646.267760] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 646.267760] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] self[:] = self._gt.wait() [ 646.267760] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.267760] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] return self._exit_event.wait() [ 646.267760] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 646.267760] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] current.throw(*self._exc) [ 646.268133] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.268133] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] result = function(*args, **kwargs) [ 646.268133] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 646.268133] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] return func(*args, **kwargs) [ 646.268133] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.268133] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] raise e [ 646.268133] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.268133] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] nwinfo = self.network_api.allocate_for_instance( [ 646.268133] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.268133] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] created_port_ids = self._update_ports_for_instance( [ 646.268133] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.268133] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] with excutils.save_and_reraise_exception(): [ 646.268133] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.268494] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] self.force_reraise() [ 646.268494] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.268494] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] raise self.value [ 646.268494] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.268494] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] updated_port = self._update_port( [ 646.268494] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.268494] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] _ensure_no_port_binding_failure(port) [ 646.268494] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.268494] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] raise exception.PortBindingFailed(port_id=port['id']) [ 646.268494] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] nova.exception.PortBindingFailed: Binding failed for port 8d185603-cceb-42c2-870a-7c89e54dff99, please check neutron logs for more information. [ 646.268494] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] [ 646.268494] env[62914]: INFO nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Terminating instance [ 646.269577] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Acquiring lock "refresh_cache-0419073e-5340-4805-a1e0-417ccf5fdd30" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.397537] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.504505] env[62914]: DEBUG nova.network.neutron [req-16f367d5-2a5c-46c5-bce6-02144e4f4d0e req-c80e6a2c-e6e1-4989-a576-e876b4cd78b8 service nova] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.601103] env[62914]: DEBUG nova.network.neutron [req-16f367d5-2a5c-46c5-bce6-02144e4f4d0e req-c80e6a2c-e6e1-4989-a576-e876b4cd78b8 service nova] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.684274] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5cf9d989-6932-4aca-9202-6e02b7d37ba5 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "f3509973-013e-45ce-87f9-357e782f26d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.372s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.094119] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-345c8103-79d2-4283-a975-910c06e967a7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.100220] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e6bfc6-b9d8-45fa-b4e4-078a693a5196 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.103884] env[62914]: DEBUG oslo_concurrency.lockutils [req-16f367d5-2a5c-46c5-bce6-02144e4f4d0e req-c80e6a2c-e6e1-4989-a576-e876b4cd78b8 service nova] Releasing lock "refresh_cache-0419073e-5340-4805-a1e0-417ccf5fdd30" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.104236] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Acquired lock "refresh_cache-0419073e-5340-4805-a1e0-417ccf5fdd30" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.104423] env[62914]: DEBUG nova.network.neutron [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 647.133155] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a040643-e57b-4c78-96d5-49b38a102f71 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.143766] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43f5bea-9e0f-4275-976e-04d76192656a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.158450] env[62914]: DEBUG nova.compute.provider_tree [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.639530] env[62914]: DEBUG oslo_concurrency.lockutils [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "34b0b81e-fc98-4019-92c3-fe2ce2fdfd67" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.639530] env[62914]: DEBUG oslo_concurrency.lockutils [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "34b0b81e-fc98-4019-92c3-fe2ce2fdfd67" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.640646] env[62914]: DEBUG oslo_concurrency.lockutils [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "34b0b81e-fc98-4019-92c3-fe2ce2fdfd67-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.640646] env[62914]: DEBUG oslo_concurrency.lockutils [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "34b0b81e-fc98-4019-92c3-fe2ce2fdfd67-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.640646] env[62914]: DEBUG oslo_concurrency.lockutils [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "34b0b81e-fc98-4019-92c3-fe2ce2fdfd67-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.642412] env[62914]: INFO nova.compute.manager [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Terminating instance [ 647.644189] env[62914]: DEBUG oslo_concurrency.lockutils [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "refresh_cache-34b0b81e-fc98-4019-92c3-fe2ce2fdfd67" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.644338] env[62914]: DEBUG oslo_concurrency.lockutils [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquired lock "refresh_cache-34b0b81e-fc98-4019-92c3-fe2ce2fdfd67" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.645235] env[62914]: DEBUG nova.network.neutron [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 647.651478] env[62914]: DEBUG nova.network.neutron [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.660778] env[62914]: DEBUG nova.scheduler.client.report [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.768384] env[62914]: DEBUG nova.network.neutron [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.007218] env[62914]: DEBUG nova.compute.manager [req-f5fa90a9-e554-4d15-b7f5-86ee7b361dbf req-8a318acb-d52a-4139-b470-e196bf240c8d service nova] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Received event network-vif-deleted-8d185603-cceb-42c2-870a-7c89e54dff99 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 648.170018] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.018s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.170018] env[62914]: ERROR nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b24a9394-2930-43a3-a75c-b52c67465668, please check neutron logs for more information. [ 648.170018] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Traceback (most recent call last): [ 648.170018] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.170018] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] self.driver.spawn(context, instance, image_meta, [ 648.170018] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 648.170018] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.170018] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.170018] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] vm_ref = self.build_virtual_machine(instance, [ 648.170792] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.170792] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.170792] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.170792] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] for vif in network_info: [ 648.170792] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.170792] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] return self._sync_wrapper(fn, *args, **kwargs) [ 648.170792] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.170792] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] self.wait() [ 648.170792] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.170792] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] self[:] = self._gt.wait() [ 648.170792] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.170792] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] return self._exit_event.wait() [ 648.170792] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 648.171143] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] current.throw(*self._exc) [ 648.171143] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.171143] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] result = function(*args, **kwargs) [ 648.171143] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 648.171143] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] return func(*args, **kwargs) [ 648.171143] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.171143] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] raise e [ 648.171143] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.171143] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] nwinfo = self.network_api.allocate_for_instance( [ 648.171143] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.171143] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] created_port_ids = self._update_ports_for_instance( [ 648.171143] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.171143] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] with excutils.save_and_reraise_exception(): [ 648.171449] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.171449] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] self.force_reraise() [ 648.171449] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.171449] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] raise self.value [ 648.171449] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.171449] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] updated_port = self._update_port( [ 648.171449] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.171449] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] _ensure_no_port_binding_failure(port) [ 648.171449] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.171449] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] raise exception.PortBindingFailed(port_id=port['id']) [ 648.171449] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] nova.exception.PortBindingFailed: Binding failed for port b24a9394-2930-43a3-a75c-b52c67465668, please check neutron logs for more information. [ 648.171449] env[62914]: ERROR nova.compute.manager [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] [ 648.171732] env[62914]: DEBUG nova.compute.utils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Binding failed for port b24a9394-2930-43a3-a75c-b52c67465668, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 648.176170] env[62914]: DEBUG nova.network.neutron [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.176170] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.904s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.180284] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Build of instance 283ffd74-3891-4b5e-bcf5-49b265978051 was re-scheduled: Binding failed for port b24a9394-2930-43a3-a75c-b52c67465668, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 648.182435] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 648.182671] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "refresh_cache-283ffd74-3891-4b5e-bcf5-49b265978051" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.182822] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquired lock "refresh_cache-283ffd74-3891-4b5e-bcf5-49b265978051" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.182979] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 648.252831] env[62914]: DEBUG nova.network.neutron [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.269272] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Releasing lock "refresh_cache-0419073e-5340-4805-a1e0-417ccf5fdd30" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.269663] env[62914]: DEBUG nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 648.269872] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 648.270152] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5b1b29c5-0922-44ba-9954-4e974058c29e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.284712] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea04a2a4-3777-4afd-8e2a-a2c4ce90ef38 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.310839] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0419073e-5340-4805-a1e0-417ccf5fdd30 could not be found. [ 648.311090] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 648.311295] env[62914]: INFO nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Took 0.04 seconds to destroy the instance on the hypervisor. [ 648.311541] env[62914]: DEBUG oslo.service.loopingcall [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 648.311822] env[62914]: DEBUG nova.compute.manager [-] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 648.311945] env[62914]: DEBUG nova.network.neutron [-] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 648.327443] env[62914]: DEBUG nova.network.neutron [-] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.707011] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.757929] env[62914]: DEBUG oslo_concurrency.lockutils [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Releasing lock "refresh_cache-34b0b81e-fc98-4019-92c3-fe2ce2fdfd67" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.758759] env[62914]: DEBUG nova.compute.manager [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 648.758759] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 648.759497] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbccae7d-7f11-4b84-8166-edc7fb8de594 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.770138] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 648.770391] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0be34299-4fc9-42a5-bdf7-23ae83380f81 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.778406] env[62914]: DEBUG oslo_vmware.api [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 648.778406] env[62914]: value = "task-1352399" [ 648.778406] env[62914]: _type = "Task" [ 648.778406] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.785982] env[62914]: DEBUG oslo_vmware.api [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352399, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.829537] env[62914]: DEBUG nova.network.neutron [-] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.884873] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.087773] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c271e77-779a-405b-bc93-c654f3858451 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.095800] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37cf4523-9ff5-45f3-b094-f0ce666ceda0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.124428] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8744191-1089-496d-9461-01017cd29b03 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.132019] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdce87de-cc69-444f-8ddd-4c1d4d59f7fd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.145507] env[62914]: DEBUG nova.compute.provider_tree [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.291114] env[62914]: DEBUG oslo_vmware.api [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352399, 'name': PowerOffVM_Task, 'duration_secs': 0.114101} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.291114] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 649.291114] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 649.291114] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d17a4046-14b9-4005-b2fd-04540e378174 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.316094] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 649.316094] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 649.316094] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Deleting the datastore file [datastore1] 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 649.316269] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-19aba096-7ea5-4c29-af98-a2a8c71dc642 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.326128] env[62914]: DEBUG oslo_vmware.api [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for the task: (returnval){ [ 649.326128] env[62914]: value = "task-1352401" [ 649.326128] env[62914]: _type = "Task" [ 649.326128] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.332220] env[62914]: INFO nova.compute.manager [-] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Took 1.02 seconds to deallocate network for instance. [ 649.334612] env[62914]: DEBUG nova.compute.claims [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 649.334782] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.338308] env[62914]: DEBUG oslo_vmware.api [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352401, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.387233] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Releasing lock "refresh_cache-283ffd74-3891-4b5e-bcf5-49b265978051" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.387481] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 649.387661] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 649.387909] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 649.417042] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.648400] env[62914]: DEBUG nova.scheduler.client.report [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.835909] env[62914]: DEBUG oslo_vmware.api [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Task: {'id': task-1352401, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.115151} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.836191] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 649.836375] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 649.836551] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 649.836720] env[62914]: INFO nova.compute.manager [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Took 1.08 seconds to destroy the instance on the hypervisor. [ 649.836962] env[62914]: DEBUG oslo.service.loopingcall [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 649.837166] env[62914]: DEBUG nova.compute.manager [-] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 649.837264] env[62914]: DEBUG nova.network.neutron [-] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 649.857069] env[62914]: DEBUG nova.network.neutron [-] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.919677] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.152949] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.977s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.153609] env[62914]: ERROR nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a3d89c65-4294-4a4e-b974-ed7f91fc188a, please check neutron logs for more information. [ 650.153609] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Traceback (most recent call last): [ 650.153609] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 650.153609] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] self.driver.spawn(context, instance, image_meta, [ 650.153609] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 650.153609] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.153609] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.153609] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] vm_ref = self.build_virtual_machine(instance, [ 650.153609] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.153609] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.153609] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.153927] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] for vif in network_info: [ 650.153927] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.153927] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] return self._sync_wrapper(fn, *args, **kwargs) [ 650.153927] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.153927] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] self.wait() [ 650.153927] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.153927] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] self[:] = self._gt.wait() [ 650.153927] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.153927] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] return self._exit_event.wait() [ 650.153927] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 650.153927] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] result = hub.switch() [ 650.153927] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 650.153927] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] return self.greenlet.switch() [ 650.154302] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.154302] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] result = function(*args, **kwargs) [ 650.154302] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 650.154302] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] return func(*args, **kwargs) [ 650.154302] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.154302] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] raise e [ 650.154302] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.154302] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] nwinfo = self.network_api.allocate_for_instance( [ 650.154302] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.154302] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] created_port_ids = self._update_ports_for_instance( [ 650.154302] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.154302] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] with excutils.save_and_reraise_exception(): [ 650.154302] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.154601] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] self.force_reraise() [ 650.154601] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.154601] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] raise self.value [ 650.154601] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.154601] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] updated_port = self._update_port( [ 650.154601] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.154601] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] _ensure_no_port_binding_failure(port) [ 650.154601] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.154601] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] raise exception.PortBindingFailed(port_id=port['id']) [ 650.154601] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] nova.exception.PortBindingFailed: Binding failed for port a3d89c65-4294-4a4e-b974-ed7f91fc188a, please check neutron logs for more information. [ 650.154601] env[62914]: ERROR nova.compute.manager [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] [ 650.154864] env[62914]: DEBUG nova.compute.utils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Binding failed for port a3d89c65-4294-4a4e-b974-ed7f91fc188a, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 650.155520] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.083s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.158438] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Build of instance bfcba5cb-d9c4-4ddd-9018-d316c13525d9 was re-scheduled: Binding failed for port a3d89c65-4294-4a4e-b974-ed7f91fc188a, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 650.158865] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 650.159098] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "refresh_cache-bfcba5cb-d9c4-4ddd-9018-d316c13525d9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.159247] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquired lock "refresh_cache-bfcba5cb-d9c4-4ddd-9018-d316c13525d9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.159402] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 650.360102] env[62914]: DEBUG nova.network.neutron [-] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.423251] env[62914]: INFO nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 283ffd74-3891-4b5e-bcf5-49b265978051] Took 1.03 seconds to deallocate network for instance. [ 650.681545] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.794571] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.862718] env[62914]: INFO nova.compute.manager [-] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Took 1.03 seconds to deallocate network for instance. [ 651.154492] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b3ae4b0-6edb-45b4-a137-2ce0d28004ef {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.163355] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0164c01b-21fc-4317-ac35-282a227fb3c9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.195492] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f49af08e-d254-436f-aabb-34b3b19cd6c3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.204608] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d14fea9-0bc7-4274-8ac5-42e2e2fd8c0b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.218575] env[62914]: DEBUG nova.compute.provider_tree [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.297217] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Releasing lock "refresh_cache-bfcba5cb-d9c4-4ddd-9018-d316c13525d9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.297443] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 651.297758] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.297818] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 651.317191] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.377493] env[62914]: DEBUG oslo_concurrency.lockutils [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.458506] env[62914]: INFO nova.scheduler.client.report [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Deleted allocations for instance 283ffd74-3891-4b5e-bcf5-49b265978051 [ 651.660839] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Acquiring lock "aa1e66d1-03fb-4340-83f1-eb79c29b6934" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.661095] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Lock "aa1e66d1-03fb-4340-83f1-eb79c29b6934" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.721815] env[62914]: DEBUG nova.scheduler.client.report [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 651.822945] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.971101] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "283ffd74-3891-4b5e-bcf5-49b265978051" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.632s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.227332] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.072s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.227950] env[62914]: ERROR nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d730aaf3-0ed0-4e80-8eee-4257ababfe57, please check neutron logs for more information. [ 652.227950] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Traceback (most recent call last): [ 652.227950] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.227950] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] self.driver.spawn(context, instance, image_meta, [ 652.227950] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 652.227950] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.227950] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.227950] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] vm_ref = self.build_virtual_machine(instance, [ 652.227950] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.227950] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.227950] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.228414] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] for vif in network_info: [ 652.228414] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.228414] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] return self._sync_wrapper(fn, *args, **kwargs) [ 652.228414] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.228414] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] self.wait() [ 652.228414] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.228414] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] self[:] = self._gt.wait() [ 652.228414] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.228414] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] return self._exit_event.wait() [ 652.228414] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 652.228414] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] current.throw(*self._exc) [ 652.228414] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.228414] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] result = function(*args, **kwargs) [ 652.228726] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 652.228726] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] return func(*args, **kwargs) [ 652.228726] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.228726] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] raise e [ 652.228726] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.228726] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] nwinfo = self.network_api.allocate_for_instance( [ 652.228726] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 652.228726] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] created_port_ids = self._update_ports_for_instance( [ 652.228726] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 652.228726] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] with excutils.save_and_reraise_exception(): [ 652.228726] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.228726] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] self.force_reraise() [ 652.228726] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.229051] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] raise self.value [ 652.229051] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 652.229051] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] updated_port = self._update_port( [ 652.229051] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.229051] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] _ensure_no_port_binding_failure(port) [ 652.229051] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.229051] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] raise exception.PortBindingFailed(port_id=port['id']) [ 652.229051] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] nova.exception.PortBindingFailed: Binding failed for port d730aaf3-0ed0-4e80-8eee-4257ababfe57, please check neutron logs for more information. [ 652.229051] env[62914]: ERROR nova.compute.manager [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] [ 652.229051] env[62914]: DEBUG nova.compute.utils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Binding failed for port d730aaf3-0ed0-4e80-8eee-4257ababfe57, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 652.229901] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.659s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.232858] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Build of instance 86784639-8ac7-4f67-be5b-05c3ab1229b7 was re-scheduled: Binding failed for port d730aaf3-0ed0-4e80-8eee-4257ababfe57, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 652.233282] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 652.233485] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquiring lock "refresh_cache-86784639-8ac7-4f67-be5b-05c3ab1229b7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.233626] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Acquired lock "refresh_cache-86784639-8ac7-4f67-be5b-05c3ab1229b7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.233802] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 652.326514] env[62914]: INFO nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: bfcba5cb-d9c4-4ddd-9018-d316c13525d9] Took 1.03 seconds to deallocate network for instance. [ 652.474101] env[62914]: DEBUG nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 652.763508] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.862752] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.999698] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.116147] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e67237-40f7-4247-a537-64667af9f0aa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.124032] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-137e37bf-3ab2-45cd-ab71-2756a89321f2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.157820] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93346f4-739d-4867-aeb8-dd4f7cc69d08 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.163283] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047a311b-83e4-4e18-b085-8abca7fecc59 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.178385] env[62914]: DEBUG nova.compute.provider_tree [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.370518] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Releasing lock "refresh_cache-86784639-8ac7-4f67-be5b-05c3ab1229b7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.370518] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 653.370518] env[62914]: DEBUG nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.370518] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 653.384976] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.688189] env[62914]: DEBUG nova.scheduler.client.report [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 653.854757] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "bfcba5cb-d9c4-4ddd-9018-d316c13525d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.420s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.886802] env[62914]: DEBUG nova.network.neutron [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.192524] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.962s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.192995] env[62914]: ERROR nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e9e80771-8137-4491-b800-397ba5b4f512, please check neutron logs for more information. [ 654.192995] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Traceback (most recent call last): [ 654.192995] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 654.192995] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] self.driver.spawn(context, instance, image_meta, [ 654.192995] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 654.192995] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.192995] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.192995] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] vm_ref = self.build_virtual_machine(instance, [ 654.192995] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.192995] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.192995] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.193528] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] for vif in network_info: [ 654.193528] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.193528] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] return self._sync_wrapper(fn, *args, **kwargs) [ 654.193528] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.193528] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] self.wait() [ 654.193528] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.193528] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] self[:] = self._gt.wait() [ 654.193528] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.193528] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] return self._exit_event.wait() [ 654.193528] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 654.193528] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] result = hub.switch() [ 654.193528] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 654.193528] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] return self.greenlet.switch() [ 654.194166] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.194166] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] result = function(*args, **kwargs) [ 654.194166] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 654.194166] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] return func(*args, **kwargs) [ 654.194166] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.194166] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] raise e [ 654.194166] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.194166] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] nwinfo = self.network_api.allocate_for_instance( [ 654.194166] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.194166] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] created_port_ids = self._update_ports_for_instance( [ 654.194166] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.194166] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] with excutils.save_and_reraise_exception(): [ 654.194166] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.194740] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] self.force_reraise() [ 654.194740] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.194740] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] raise self.value [ 654.194740] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.194740] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] updated_port = self._update_port( [ 654.194740] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.194740] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] _ensure_no_port_binding_failure(port) [ 654.194740] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.194740] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] raise exception.PortBindingFailed(port_id=port['id']) [ 654.194740] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] nova.exception.PortBindingFailed: Binding failed for port e9e80771-8137-4491-b800-397ba5b4f512, please check neutron logs for more information. [ 654.194740] env[62914]: ERROR nova.compute.manager [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] [ 654.195276] env[62914]: DEBUG nova.compute.utils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Binding failed for port e9e80771-8137-4491-b800-397ba5b4f512, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 654.195276] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.116s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.198160] env[62914]: DEBUG nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Build of instance 8bda7658-2b86-4869-89e2-b10d7119ae18 was re-scheduled: Binding failed for port e9e80771-8137-4491-b800-397ba5b4f512, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 654.198604] env[62914]: DEBUG nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 654.198828] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Acquiring lock "refresh_cache-8bda7658-2b86-4869-89e2-b10d7119ae18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.198973] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Acquired lock "refresh_cache-8bda7658-2b86-4869-89e2-b10d7119ae18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.199143] env[62914]: DEBUG nova.network.neutron [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 654.360712] env[62914]: DEBUG nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 654.389227] env[62914]: INFO nova.compute.manager [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] [instance: 86784639-8ac7-4f67-be5b-05c3ab1229b7] Took 1.02 seconds to deallocate network for instance. [ 654.719634] env[62914]: DEBUG nova.network.neutron [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.823485] env[62914]: DEBUG nova.network.neutron [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.884198] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.081987] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4f8486c-f75c-4189-ac08-6ef6b2d620d6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.089697] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d8a47f-e29f-4482-b99a-8c4b354d1e3b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.119844] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef2f691b-50d4-4ed4-88e5-23d7f464d5d7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.127369] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5be5ca-5c1c-4290-bc48-9f6e08c2e837 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.140665] env[62914]: DEBUG nova.compute.provider_tree [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.326121] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Releasing lock "refresh_cache-8bda7658-2b86-4869-89e2-b10d7119ae18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.326360] env[62914]: DEBUG nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 655.326542] env[62914]: DEBUG nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.326710] env[62914]: DEBUG nova.network.neutron [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 655.346121] env[62914]: DEBUG nova.network.neutron [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.643845] env[62914]: DEBUG nova.scheduler.client.report [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.848629] env[62914]: DEBUG nova.network.neutron [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.148963] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.954s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.149681] env[62914]: ERROR nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5c57f21b-7d47-4d5b-96fe-2bbc4d523a5f, please check neutron logs for more information. [ 656.149681] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Traceback (most recent call last): [ 656.149681] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.149681] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] self.driver.spawn(context, instance, image_meta, [ 656.149681] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 656.149681] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.149681] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.149681] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] vm_ref = self.build_virtual_machine(instance, [ 656.149681] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.149681] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.149681] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.150030] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] for vif in network_info: [ 656.150030] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.150030] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] return self._sync_wrapper(fn, *args, **kwargs) [ 656.150030] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.150030] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] self.wait() [ 656.150030] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.150030] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] self[:] = self._gt.wait() [ 656.150030] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.150030] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] return self._exit_event.wait() [ 656.150030] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 656.150030] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] current.throw(*self._exc) [ 656.150030] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.150030] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] result = function(*args, **kwargs) [ 656.150343] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.150343] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] return func(*args, **kwargs) [ 656.150343] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.150343] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] raise e [ 656.150343] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.150343] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] nwinfo = self.network_api.allocate_for_instance( [ 656.150343] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.150343] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] created_port_ids = self._update_ports_for_instance( [ 656.150343] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.150343] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] with excutils.save_and_reraise_exception(): [ 656.150343] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.150343] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] self.force_reraise() [ 656.150343] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.150653] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] raise self.value [ 656.150653] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.150653] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] updated_port = self._update_port( [ 656.150653] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.150653] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] _ensure_no_port_binding_failure(port) [ 656.150653] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.150653] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] raise exception.PortBindingFailed(port_id=port['id']) [ 656.150653] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] nova.exception.PortBindingFailed: Binding failed for port 5c57f21b-7d47-4d5b-96fe-2bbc4d523a5f, please check neutron logs for more information. [ 656.150653] env[62914]: ERROR nova.compute.manager [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] [ 656.150653] env[62914]: DEBUG nova.compute.utils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Binding failed for port 5c57f21b-7d47-4d5b-96fe-2bbc4d523a5f, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 656.151628] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.636s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.153223] env[62914]: INFO nova.compute.claims [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 656.156048] env[62914]: DEBUG nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Build of instance 15938dfe-a697-450b-8c88-ff4ef703ee1f was re-scheduled: Binding failed for port 5c57f21b-7d47-4d5b-96fe-2bbc4d523a5f, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 656.156463] env[62914]: DEBUG nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 656.156695] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Acquiring lock "refresh_cache-15938dfe-a697-450b-8c88-ff4ef703ee1f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.156876] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Acquired lock "refresh_cache-15938dfe-a697-450b-8c88-ff4ef703ee1f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.157039] env[62914]: DEBUG nova.network.neutron [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.352489] env[62914]: INFO nova.compute.manager [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] [instance: 8bda7658-2b86-4869-89e2-b10d7119ae18] Took 1.03 seconds to deallocate network for instance. [ 656.426533] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ca5ba2c-e15c-42a6-82af-9f8413955f6e tempest-ListServersNegativeTestJSON-1002809678 tempest-ListServersNegativeTestJSON-1002809678-project-member] Lock "86784639-8ac7-4f67-be5b-05c3ab1229b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.937s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.427231] env[62914]: Traceback (most recent call last): [ 656.427297] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.427297] env[62914]: self.driver.spawn(context, instance, image_meta, [ 656.427297] env[62914]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 656.427297] env[62914]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.427297] env[62914]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.427297] env[62914]: vm_ref = self.build_virtual_machine(instance, [ 656.427297] env[62914]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.427297] env[62914]: vif_infos = vmwarevif.get_vif_info(self._session, [ 656.427297] env[62914]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.427297] env[62914]: for vif in network_info: [ 656.427297] env[62914]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.427297] env[62914]: return self._sync_wrapper(fn, *args, **kwargs) [ 656.427297] env[62914]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.427297] env[62914]: self.wait() [ 656.427297] env[62914]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.427297] env[62914]: self[:] = self._gt.wait() [ 656.427297] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.427297] env[62914]: return self._exit_event.wait() [ 656.427297] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 656.427297] env[62914]: current.throw(*self._exc) [ 656.427297] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.427297] env[62914]: result = function(*args, **kwargs) [ 656.427297] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.427297] env[62914]: return func(*args, **kwargs) [ 656.427297] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.427297] env[62914]: raise e [ 656.427297] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.427297] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 656.427297] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.427297] env[62914]: created_port_ids = self._update_ports_for_instance( [ 656.427297] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.427297] env[62914]: with excutils.save_and_reraise_exception(): [ 656.428119] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.428119] env[62914]: self.force_reraise() [ 656.428119] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.428119] env[62914]: raise self.value [ 656.428119] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.428119] env[62914]: updated_port = self._update_port( [ 656.428119] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.428119] env[62914]: _ensure_no_port_binding_failure(port) [ 656.428119] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.428119] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 656.428119] env[62914]: nova.exception.PortBindingFailed: Binding failed for port d730aaf3-0ed0-4e80-8eee-4257ababfe57, please check neutron logs for more information. [ 656.428119] env[62914]: During handling of the above exception, another exception occurred: [ 656.428119] env[62914]: Traceback (most recent call last): [ 656.428119] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2456, in _do_build_and_run_instance [ 656.428119] env[62914]: self._build_and_run_instance(context, instance, image, [ 656.428119] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2748, in _build_and_run_instance [ 656.428119] env[62914]: raise exception.RescheduledException( [ 656.428119] env[62914]: nova.exception.RescheduledException: Build of instance 86784639-8ac7-4f67-be5b-05c3ab1229b7 was re-scheduled: Binding failed for port d730aaf3-0ed0-4e80-8eee-4257ababfe57, please check neutron logs for more information. [ 656.428119] env[62914]: During handling of the above exception, another exception occurred: [ 656.428119] env[62914]: Traceback (most recent call last): [ 656.428119] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 656.428119] env[62914]: func(*args, **kwargs) [ 656.428119] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.428119] env[62914]: return func(*args, **kwargs) [ 656.428119] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 656.428119] env[62914]: return f(*args, **kwargs) [ 656.428119] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2347, in _locked_do_build_and_run_instance [ 656.428119] env[62914]: result = self._do_build_and_run_instance(*args, **kwargs) [ 656.428884] env[62914]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 656.428884] env[62914]: with excutils.save_and_reraise_exception(): [ 656.428884] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.428884] env[62914]: self.force_reraise() [ 656.428884] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.428884] env[62914]: raise self.value [ 656.428884] env[62914]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 656.428884] env[62914]: return f(self, context, *args, **kw) [ 656.428884] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 656.428884] env[62914]: with excutils.save_and_reraise_exception(): [ 656.428884] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.428884] env[62914]: self.force_reraise() [ 656.428884] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.428884] env[62914]: raise self.value [ 656.428884] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 656.428884] env[62914]: return function(self, context, *args, **kwargs) [ 656.428884] env[62914]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 656.428884] env[62914]: return function(self, context, *args, **kwargs) [ 656.428884] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 656.428884] env[62914]: return function(self, context, *args, **kwargs) [ 656.428884] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2491, in _do_build_and_run_instance [ 656.428884] env[62914]: instance.save() [ 656.428884] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 656.428884] env[62914]: updates, result = self.indirection_api.object_action( [ 656.428884] env[62914]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 656.428884] env[62914]: return cctxt.call(context, 'object_action', objinst=objinst, [ 656.428884] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 656.428884] env[62914]: result = self.transport._send( [ 656.428884] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 656.428884] env[62914]: return self._driver.send(target, ctxt, message, [ 656.429692] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 656.429692] env[62914]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 656.429692] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 656.429692] env[62914]: raise result [ 656.429692] env[62914]: nova.exception_Remote.InstanceNotFound_Remote: Instance 86784639-8ac7-4f67-be5b-05c3ab1229b7 could not be found. [ 656.429692] env[62914]: Traceback (most recent call last): [ 656.429692] env[62914]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 656.429692] env[62914]: return getattr(target, method)(*args, **kwargs) [ 656.429692] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 656.429692] env[62914]: return fn(self, *args, **kwargs) [ 656.429692] env[62914]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 656.429692] env[62914]: old_ref, inst_ref = db.instance_update_and_get_original( [ 656.429692] env[62914]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 656.429692] env[62914]: return f(*args, **kwargs) [ 656.429692] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 656.429692] env[62914]: with excutils.save_and_reraise_exception() as ectxt: [ 656.429692] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.429692] env[62914]: self.force_reraise() [ 656.429692] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.429692] env[62914]: raise self.value [ 656.429692] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 656.429692] env[62914]: return f(*args, **kwargs) [ 656.429692] env[62914]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 656.429692] env[62914]: return f(context, *args, **kwargs) [ 656.429692] env[62914]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 656.429692] env[62914]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 656.429692] env[62914]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 656.429692] env[62914]: raise exception.InstanceNotFound(instance_id=uuid) [ 656.429692] env[62914]: nova.exception.InstanceNotFound: Instance 86784639-8ac7-4f67-be5b-05c3ab1229b7 could not be found. [ 656.677814] env[62914]: DEBUG nova.network.neutron [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.759681] env[62914]: DEBUG nova.network.neutron [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.931834] env[62914]: DEBUG nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 657.262179] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Releasing lock "refresh_cache-15938dfe-a697-450b-8c88-ff4ef703ee1f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.262570] env[62914]: DEBUG nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 657.262788] env[62914]: DEBUG nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.262958] env[62914]: DEBUG nova.network.neutron [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 657.279171] env[62914]: DEBUG nova.network.neutron [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.375490] env[62914]: INFO nova.scheduler.client.report [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Deleted allocations for instance 8bda7658-2b86-4869-89e2-b10d7119ae18 [ 657.452962] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.484392] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84cb0d57-18b4-494a-81fb-c0fb4301d18a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.493685] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce8b0ee-73e7-415c-9faa-34eb2c1e373f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.520823] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b357c956-4888-4e5a-9b16-c1e1a7c4e9eb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.528867] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab8da71-f5b0-4d27-95ef-78671ae33be2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.542905] env[62914]: DEBUG nova.compute.provider_tree [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.781583] env[62914]: DEBUG nova.network.neutron [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.885592] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dac31136-76d5-43ee-8566-875efb51d6dd tempest-VolumesAdminNegativeTest-831555146 tempest-VolumesAdminNegativeTest-831555146-project-member] Lock "8bda7658-2b86-4869-89e2-b10d7119ae18" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.907s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.045631] env[62914]: DEBUG nova.scheduler.client.report [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.284034] env[62914]: INFO nova.compute.manager [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] [instance: 15938dfe-a697-450b-8c88-ff4ef703ee1f] Took 1.02 seconds to deallocate network for instance. [ 658.387826] env[62914]: DEBUG nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 658.551922] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.552251] env[62914]: DEBUG nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 658.554931] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.993s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.556438] env[62914]: INFO nova.compute.claims [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 658.913082] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.064481] env[62914]: DEBUG nova.compute.utils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 659.065970] env[62914]: DEBUG nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 659.066149] env[62914]: DEBUG nova.network.neutron [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 659.106944] env[62914]: DEBUG nova.policy [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3727f2fcfdc748c69ca19d03beb6255f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dde6a268571045beb738fed3d1c2bab2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 659.314044] env[62914]: INFO nova.scheduler.client.report [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Deleted allocations for instance 15938dfe-a697-450b-8c88-ff4ef703ee1f [ 659.578084] env[62914]: DEBUG nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 659.592032] env[62914]: DEBUG nova.network.neutron [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Successfully created port: 7fef89a0-fac4-420f-8c07-a91a671d5de9 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 659.822816] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f625413d-0196-4079-8f64-099493c7441e tempest-ServerActionsTestJSON-1426609420 tempest-ServerActionsTestJSON-1426609420-project-member] Lock "15938dfe-a697-450b-8c88-ff4ef703ee1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.845s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.974429] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8582ba73-99c3-45c6-b2ff-ff47e690f088 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.987261] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba8377d6-4a07-4e9f-9746-0fbc5239c478 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.020792] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c828a45d-28ea-4ca6-a0be-98a2b2929d66 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.028912] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be7ab29-ace1-4f0a-bf1d-0e999b9626a5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.046320] env[62914]: DEBUG nova.compute.provider_tree [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.325091] env[62914]: DEBUG nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 660.550192] env[62914]: DEBUG nova.scheduler.client.report [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 660.588148] env[62914]: DEBUG nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 660.623341] env[62914]: DEBUG nova.virt.hardware [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 660.623683] env[62914]: DEBUG nova.virt.hardware [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 660.623922] env[62914]: DEBUG nova.virt.hardware [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 660.624171] env[62914]: DEBUG nova.virt.hardware [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 660.624449] env[62914]: DEBUG nova.virt.hardware [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 660.624561] env[62914]: DEBUG nova.virt.hardware [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 660.624873] env[62914]: DEBUG nova.virt.hardware [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 660.625103] env[62914]: DEBUG nova.virt.hardware [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 660.625316] env[62914]: DEBUG nova.virt.hardware [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 660.625670] env[62914]: DEBUG nova.virt.hardware [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 660.625781] env[62914]: DEBUG nova.virt.hardware [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 660.626741] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb91445d-7c10-4dd8-baaa-5c70ca289baa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.638574] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b80e5af-bd2d-4e88-b296-b2f6054aa64f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.658954] env[62914]: DEBUG nova.compute.manager [req-142461a1-9603-4a56-b846-1c07aa3a3da6 req-2ce9308b-7d9b-411d-b4ef-a88b58017a2d service nova] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Received event network-changed-7fef89a0-fac4-420f-8c07-a91a671d5de9 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 660.659202] env[62914]: DEBUG nova.compute.manager [req-142461a1-9603-4a56-b846-1c07aa3a3da6 req-2ce9308b-7d9b-411d-b4ef-a88b58017a2d service nova] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Refreshing instance network info cache due to event network-changed-7fef89a0-fac4-420f-8c07-a91a671d5de9. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 660.659445] env[62914]: DEBUG oslo_concurrency.lockutils [req-142461a1-9603-4a56-b846-1c07aa3a3da6 req-2ce9308b-7d9b-411d-b4ef-a88b58017a2d service nova] Acquiring lock "refresh_cache-c10d840e-ce05-4af3-ba2d-6214eefd8783" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.659930] env[62914]: DEBUG oslo_concurrency.lockutils [req-142461a1-9603-4a56-b846-1c07aa3a3da6 req-2ce9308b-7d9b-411d-b4ef-a88b58017a2d service nova] Acquired lock "refresh_cache-c10d840e-ce05-4af3-ba2d-6214eefd8783" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.659930] env[62914]: DEBUG nova.network.neutron [req-142461a1-9603-4a56-b846-1c07aa3a3da6 req-2ce9308b-7d9b-411d-b4ef-a88b58017a2d service nova] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Refreshing network info cache for port 7fef89a0-fac4-420f-8c07-a91a671d5de9 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 660.854664] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.867936] env[62914]: ERROR nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7fef89a0-fac4-420f-8c07-a91a671d5de9, please check neutron logs for more information. [ 660.867936] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 660.867936] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.867936] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 660.867936] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.867936] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 660.867936] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.867936] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 660.867936] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.867936] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 660.867936] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.867936] env[62914]: ERROR nova.compute.manager raise self.value [ 660.867936] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.867936] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 660.867936] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.867936] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 660.868400] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.868400] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 660.868400] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7fef89a0-fac4-420f-8c07-a91a671d5de9, please check neutron logs for more information. [ 660.868400] env[62914]: ERROR nova.compute.manager [ 660.868400] env[62914]: Traceback (most recent call last): [ 660.868400] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 660.868400] env[62914]: listener.cb(fileno) [ 660.868400] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.868400] env[62914]: result = function(*args, **kwargs) [ 660.868400] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 660.868400] env[62914]: return func(*args, **kwargs) [ 660.868400] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.868400] env[62914]: raise e [ 660.868400] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.868400] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 660.868400] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.868400] env[62914]: created_port_ids = self._update_ports_for_instance( [ 660.868400] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.868400] env[62914]: with excutils.save_and_reraise_exception(): [ 660.868400] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.868400] env[62914]: self.force_reraise() [ 660.868400] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.868400] env[62914]: raise self.value [ 660.868400] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.868400] env[62914]: updated_port = self._update_port( [ 660.868400] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.868400] env[62914]: _ensure_no_port_binding_failure(port) [ 660.868400] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.868400] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 660.869086] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 7fef89a0-fac4-420f-8c07-a91a671d5de9, please check neutron logs for more information. [ 660.869086] env[62914]: Removing descriptor: 19 [ 660.869086] env[62914]: ERROR nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7fef89a0-fac4-420f-8c07-a91a671d5de9, please check neutron logs for more information. [ 660.869086] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Traceback (most recent call last): [ 660.869086] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 660.869086] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] yield resources [ 660.869086] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.869086] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] self.driver.spawn(context, instance, image_meta, [ 660.869086] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 660.869086] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.869086] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.869086] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] vm_ref = self.build_virtual_machine(instance, [ 660.869385] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.869385] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.869385] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.869385] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] for vif in network_info: [ 660.869385] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.869385] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] return self._sync_wrapper(fn, *args, **kwargs) [ 660.869385] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.869385] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] self.wait() [ 660.869385] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.869385] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] self[:] = self._gt.wait() [ 660.869385] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.869385] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] return self._exit_event.wait() [ 660.869385] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.869725] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] result = hub.switch() [ 660.869725] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.869725] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] return self.greenlet.switch() [ 660.869725] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.869725] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] result = function(*args, **kwargs) [ 660.869725] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 660.869725] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] return func(*args, **kwargs) [ 660.869725] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.869725] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] raise e [ 660.869725] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.869725] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] nwinfo = self.network_api.allocate_for_instance( [ 660.869725] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.869725] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] created_port_ids = self._update_ports_for_instance( [ 660.870047] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.870047] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] with excutils.save_and_reraise_exception(): [ 660.870047] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.870047] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] self.force_reraise() [ 660.870047] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.870047] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] raise self.value [ 660.870047] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.870047] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] updated_port = self._update_port( [ 660.870047] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.870047] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] _ensure_no_port_binding_failure(port) [ 660.870047] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.870047] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] raise exception.PortBindingFailed(port_id=port['id']) [ 660.870336] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] nova.exception.PortBindingFailed: Binding failed for port 7fef89a0-fac4-420f-8c07-a91a671d5de9, please check neutron logs for more information. [ 660.870336] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] [ 660.870336] env[62914]: INFO nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Terminating instance [ 660.871387] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Acquiring lock "refresh_cache-c10d840e-ce05-4af3-ba2d-6214eefd8783" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.059017] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.059017] env[62914]: DEBUG nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 661.063882] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.416s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.175769] env[62914]: DEBUG nova.network.neutron [req-142461a1-9603-4a56-b846-1c07aa3a3da6 req-2ce9308b-7d9b-411d-b4ef-a88b58017a2d service nova] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.264685] env[62914]: DEBUG nova.network.neutron [req-142461a1-9603-4a56-b846-1c07aa3a3da6 req-2ce9308b-7d9b-411d-b4ef-a88b58017a2d service nova] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.571025] env[62914]: DEBUG nova.compute.utils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 661.574244] env[62914]: DEBUG nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 661.574761] env[62914]: DEBUG nova.network.neutron [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 661.651508] env[62914]: DEBUG nova.policy [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce06463e715b4975bc43b14be64ef9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '619be7b3bf1445b68ba7adefc98f8782', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 661.770746] env[62914]: DEBUG oslo_concurrency.lockutils [req-142461a1-9603-4a56-b846-1c07aa3a3da6 req-2ce9308b-7d9b-411d-b4ef-a88b58017a2d service nova] Releasing lock "refresh_cache-c10d840e-ce05-4af3-ba2d-6214eefd8783" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.771183] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Acquired lock "refresh_cache-c10d840e-ce05-4af3-ba2d-6214eefd8783" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.771486] env[62914]: DEBUG nova.network.neutron [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 662.017755] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60efb2cb-b094-4c82-b5d0-9f31010aa7c5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.026239] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b76b41-73eb-45b7-95a1-7f74fae0f61c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.061713] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ef217c-027c-451f-8c69-8174110db244 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.069692] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8183e5f-5f25-406c-bfba-5f65b8baee3c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.075721] env[62914]: DEBUG nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 662.087740] env[62914]: DEBUG nova.compute.provider_tree [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.090484] env[62914]: DEBUG nova.network.neutron [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Successfully created port: 83e791da-9e3a-4c0f-9454-6662da4e8798 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 662.294568] env[62914]: DEBUG nova.network.neutron [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.435293] env[62914]: DEBUG nova.network.neutron [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.593265] env[62914]: DEBUG nova.scheduler.client.report [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.712071] env[62914]: DEBUG nova.compute.manager [req-f0855d32-906d-4b3d-b901-9548d99b1d4d req-acb6ad6c-4a79-47d6-ada2-8d4da9c76142 service nova] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Received event network-vif-deleted-7fef89a0-fac4-420f-8c07-a91a671d5de9 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 662.938569] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Releasing lock "refresh_cache-c10d840e-ce05-4af3-ba2d-6214eefd8783" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.939081] env[62914]: DEBUG nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 662.939189] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 662.939487] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c97e39c-e4d4-4a1a-ad78-618476550a5c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.949168] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87a93aa-173e-4dc9-8090-03be78716a54 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.976266] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c10d840e-ce05-4af3-ba2d-6214eefd8783 could not be found. [ 662.976590] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 662.976668] env[62914]: INFO nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Took 0.04 seconds to destroy the instance on the hypervisor. [ 662.976910] env[62914]: DEBUG oslo.service.loopingcall [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 662.977170] env[62914]: DEBUG nova.compute.manager [-] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.977265] env[62914]: DEBUG nova.network.neutron [-] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 663.006130] env[62914]: DEBUG nova.network.neutron [-] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.094058] env[62914]: DEBUG nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 663.107024] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.040s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.107024] env[62914]: ERROR nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3589f50e-46e4-4a1f-8b9d-51955b48102d, please check neutron logs for more information. [ 663.107024] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Traceback (most recent call last): [ 663.107024] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.107024] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] self.driver.spawn(context, instance, image_meta, [ 663.107024] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 663.107024] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.107024] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.107024] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] vm_ref = self.build_virtual_machine(instance, [ 663.107360] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.107360] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.107360] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.107360] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] for vif in network_info: [ 663.107360] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.107360] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] return self._sync_wrapper(fn, *args, **kwargs) [ 663.107360] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.107360] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] self.wait() [ 663.107360] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.107360] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] self[:] = self._gt.wait() [ 663.107360] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.107360] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] return self._exit_event.wait() [ 663.107360] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 663.107771] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] result = hub.switch() [ 663.107771] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 663.107771] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] return self.greenlet.switch() [ 663.107771] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.107771] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] result = function(*args, **kwargs) [ 663.107771] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 663.107771] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] return func(*args, **kwargs) [ 663.107771] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.107771] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] raise e [ 663.107771] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.107771] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] nwinfo = self.network_api.allocate_for_instance( [ 663.107771] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.107771] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] created_port_ids = self._update_ports_for_instance( [ 663.108118] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.108118] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] with excutils.save_and_reraise_exception(): [ 663.108118] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.108118] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] self.force_reraise() [ 663.108118] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.108118] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] raise self.value [ 663.108118] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.108118] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] updated_port = self._update_port( [ 663.108118] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.108118] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] _ensure_no_port_binding_failure(port) [ 663.108118] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.108118] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] raise exception.PortBindingFailed(port_id=port['id']) [ 663.108541] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] nova.exception.PortBindingFailed: Binding failed for port 3589f50e-46e4-4a1f-8b9d-51955b48102d, please check neutron logs for more information. [ 663.108541] env[62914]: ERROR nova.compute.manager [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] [ 663.108541] env[62914]: DEBUG nova.compute.utils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Binding failed for port 3589f50e-46e4-4a1f-8b9d-51955b48102d, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 663.112365] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.893s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.114167] env[62914]: DEBUG nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Build of instance 91328e42-ba3c-4d63-a805-973c2bfa2afb was re-scheduled: Binding failed for port 3589f50e-46e4-4a1f-8b9d-51955b48102d, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 663.114609] env[62914]: DEBUG nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 663.114830] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Acquiring lock "refresh_cache-91328e42-ba3c-4d63-a805-973c2bfa2afb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.115013] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Acquired lock "refresh_cache-91328e42-ba3c-4d63-a805-973c2bfa2afb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.116586] env[62914]: DEBUG nova.network.neutron [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 663.136852] env[62914]: DEBUG nova.virt.hardware [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 663.139455] env[62914]: DEBUG nova.virt.hardware [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 663.139455] env[62914]: DEBUG nova.virt.hardware [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 663.139455] env[62914]: DEBUG nova.virt.hardware [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 663.139455] env[62914]: DEBUG nova.virt.hardware [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 663.139455] env[62914]: DEBUG nova.virt.hardware [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 663.139681] env[62914]: DEBUG nova.virt.hardware [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 663.139681] env[62914]: DEBUG nova.virt.hardware [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 663.139681] env[62914]: DEBUG nova.virt.hardware [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 663.139681] env[62914]: DEBUG nova.virt.hardware [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 663.139681] env[62914]: DEBUG nova.virt.hardware [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 663.139847] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d2e324-42ba-4ddb-8242-5823a4f5e40f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.148971] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00234127-925f-42ee-b692-856853c98ce0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.362050] env[62914]: ERROR nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 83e791da-9e3a-4c0f-9454-6662da4e8798, please check neutron logs for more information. [ 663.362050] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 663.362050] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.362050] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 663.362050] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.362050] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 663.362050] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.362050] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 663.362050] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.362050] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 663.362050] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.362050] env[62914]: ERROR nova.compute.manager raise self.value [ 663.362050] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.362050] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 663.362050] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.362050] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 663.362542] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.362542] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 663.362542] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 83e791da-9e3a-4c0f-9454-6662da4e8798, please check neutron logs for more information. [ 663.362542] env[62914]: ERROR nova.compute.manager [ 663.362542] env[62914]: Traceback (most recent call last): [ 663.362542] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 663.362542] env[62914]: listener.cb(fileno) [ 663.362542] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.362542] env[62914]: result = function(*args, **kwargs) [ 663.362542] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 663.362542] env[62914]: return func(*args, **kwargs) [ 663.362542] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.362542] env[62914]: raise e [ 663.362542] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.362542] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 663.362542] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.362542] env[62914]: created_port_ids = self._update_ports_for_instance( [ 663.362542] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.362542] env[62914]: with excutils.save_and_reraise_exception(): [ 663.362542] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.362542] env[62914]: self.force_reraise() [ 663.362542] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.362542] env[62914]: raise self.value [ 663.362542] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.362542] env[62914]: updated_port = self._update_port( [ 663.362542] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.362542] env[62914]: _ensure_no_port_binding_failure(port) [ 663.362542] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.362542] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 663.363376] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 83e791da-9e3a-4c0f-9454-6662da4e8798, please check neutron logs for more information. [ 663.363376] env[62914]: Removing descriptor: 19 [ 663.363376] env[62914]: ERROR nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 83e791da-9e3a-4c0f-9454-6662da4e8798, please check neutron logs for more information. [ 663.363376] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Traceback (most recent call last): [ 663.363376] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 663.363376] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] yield resources [ 663.363376] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.363376] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] self.driver.spawn(context, instance, image_meta, [ 663.363376] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 663.363376] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.363376] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.363376] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] vm_ref = self.build_virtual_machine(instance, [ 663.363762] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.363762] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.363762] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.363762] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] for vif in network_info: [ 663.363762] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.363762] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] return self._sync_wrapper(fn, *args, **kwargs) [ 663.363762] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.363762] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] self.wait() [ 663.363762] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.363762] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] self[:] = self._gt.wait() [ 663.363762] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.363762] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] return self._exit_event.wait() [ 663.363762] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 663.364097] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] result = hub.switch() [ 663.364097] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 663.364097] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] return self.greenlet.switch() [ 663.364097] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.364097] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] result = function(*args, **kwargs) [ 663.364097] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 663.364097] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] return func(*args, **kwargs) [ 663.364097] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.364097] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] raise e [ 663.364097] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.364097] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] nwinfo = self.network_api.allocate_for_instance( [ 663.364097] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.364097] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] created_port_ids = self._update_ports_for_instance( [ 663.364444] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.364444] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] with excutils.save_and_reraise_exception(): [ 663.364444] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.364444] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] self.force_reraise() [ 663.364444] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.364444] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] raise self.value [ 663.364444] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.364444] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] updated_port = self._update_port( [ 663.364444] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.364444] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] _ensure_no_port_binding_failure(port) [ 663.364444] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.364444] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] raise exception.PortBindingFailed(port_id=port['id']) [ 663.364789] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] nova.exception.PortBindingFailed: Binding failed for port 83e791da-9e3a-4c0f-9454-6662da4e8798, please check neutron logs for more information. [ 663.364789] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] [ 663.364789] env[62914]: INFO nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Terminating instance [ 663.364789] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "refresh_cache-c5aaadb0-f9a1-4af9-b882-86d847185a19" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.364789] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "refresh_cache-c5aaadb0-f9a1-4af9-b882-86d847185a19" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.364789] env[62914]: DEBUG nova.network.neutron [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 663.508927] env[62914]: DEBUG nova.network.neutron [-] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.648404] env[62914]: DEBUG nova.network.neutron [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.730481] env[62914]: DEBUG nova.network.neutron [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.885233] env[62914]: DEBUG nova.network.neutron [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.013573] env[62914]: INFO nova.compute.manager [-] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Took 1.03 seconds to deallocate network for instance. [ 664.014755] env[62914]: DEBUG nova.compute.claims [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 664.014755] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.045568] env[62914]: DEBUG nova.network.neutron [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.153914] env[62914]: WARNING nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 664.236858] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Releasing lock "refresh_cache-91328e42-ba3c-4d63-a805-973c2bfa2afb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.237130] env[62914]: DEBUG nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 664.239150] env[62914]: DEBUG nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.239150] env[62914]: DEBUG nova.network.neutron [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 664.258065] env[62914]: DEBUG nova.network.neutron [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.547299] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "refresh_cache-c5aaadb0-f9a1-4af9-b882-86d847185a19" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.547715] env[62914]: DEBUG nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 664.548177] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 664.548543] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-466f7923-19ed-4f38-88c2-95dbe31a1985 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.558316] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767f1e4b-8ef9-4b8c-90fa-e81235987738 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.582449] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c5aaadb0-f9a1-4af9-b882-86d847185a19 could not be found. [ 664.582672] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 664.582847] env[62914]: INFO nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Took 0.03 seconds to destroy the instance on the hypervisor. [ 664.583105] env[62914]: DEBUG oslo.service.loopingcall [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 664.583345] env[62914]: DEBUG nova.compute.manager [-] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.583436] env[62914]: DEBUG nova.network.neutron [-] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 664.617116] env[62914]: DEBUG nova.network.neutron [-] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.658663] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 91328e42-ba3c-4d63-a805-973c2bfa2afb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 664.659192] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 0419073e-5340-4805-a1e0-417ccf5fdd30 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 664.659192] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance c10d840e-ce05-4af3-ba2d-6214eefd8783 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 664.659290] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance c5aaadb0-f9a1-4af9-b882-86d847185a19 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 664.743017] env[62914]: DEBUG nova.compute.manager [req-1513e953-a1af-415e-ab1d-085747626995 req-394dd28f-874f-45a9-8143-4068bfd3cfc0 service nova] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Received event network-changed-83e791da-9e3a-4c0f-9454-6662da4e8798 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 664.743017] env[62914]: DEBUG nova.compute.manager [req-1513e953-a1af-415e-ab1d-085747626995 req-394dd28f-874f-45a9-8143-4068bfd3cfc0 service nova] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Refreshing instance network info cache due to event network-changed-83e791da-9e3a-4c0f-9454-6662da4e8798. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 664.743017] env[62914]: DEBUG oslo_concurrency.lockutils [req-1513e953-a1af-415e-ab1d-085747626995 req-394dd28f-874f-45a9-8143-4068bfd3cfc0 service nova] Acquiring lock "refresh_cache-c5aaadb0-f9a1-4af9-b882-86d847185a19" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.743017] env[62914]: DEBUG oslo_concurrency.lockutils [req-1513e953-a1af-415e-ab1d-085747626995 req-394dd28f-874f-45a9-8143-4068bfd3cfc0 service nova] Acquired lock "refresh_cache-c5aaadb0-f9a1-4af9-b882-86d847185a19" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.743017] env[62914]: DEBUG nova.network.neutron [req-1513e953-a1af-415e-ab1d-085747626995 req-394dd28f-874f-45a9-8143-4068bfd3cfc0 service nova] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Refreshing network info cache for port 83e791da-9e3a-4c0f-9454-6662da4e8798 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 664.762025] env[62914]: DEBUG nova.network.neutron [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.120250] env[62914]: DEBUG nova.network.neutron [-] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.163325] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.263793] env[62914]: INFO nova.compute.manager [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 91328e42-ba3c-4d63-a805-973c2bfa2afb] Took 1.03 seconds to deallocate network for instance. [ 665.269527] env[62914]: DEBUG nova.network.neutron [req-1513e953-a1af-415e-ab1d-085747626995 req-394dd28f-874f-45a9-8143-4068bfd3cfc0 service nova] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.441114] env[62914]: DEBUG nova.network.neutron [req-1513e953-a1af-415e-ab1d-085747626995 req-394dd28f-874f-45a9-8143-4068bfd3cfc0 service nova] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.624149] env[62914]: INFO nova.compute.manager [-] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Took 1.04 seconds to deallocate network for instance. [ 665.628997] env[62914]: DEBUG nova.compute.claims [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 665.629237] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.666787] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 294e3321-e826-4a56-bbd0-893261d65283 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.943298] env[62914]: DEBUG oslo_concurrency.lockutils [req-1513e953-a1af-415e-ab1d-085747626995 req-394dd28f-874f-45a9-8143-4068bfd3cfc0 service nova] Releasing lock "refresh_cache-c5aaadb0-f9a1-4af9-b882-86d847185a19" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.945364] env[62914]: DEBUG nova.compute.manager [req-1513e953-a1af-415e-ab1d-085747626995 req-394dd28f-874f-45a9-8143-4068bfd3cfc0 service nova] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Received event network-vif-deleted-83e791da-9e3a-4c0f-9454-6662da4e8798 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 666.170133] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 0ac803ee-6303-463f-91e8-8263c0219b0a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.326020] env[62914]: INFO nova.scheduler.client.report [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Deleted allocations for instance 91328e42-ba3c-4d63-a805-973c2bfa2afb [ 666.675106] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 3ce62ec7-6274-441f-989b-6f27a64b0c18 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.832036] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1bcd40b6-f7be-4ae5-a22f-0045453523a7 tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Lock "91328e42-ba3c-4d63-a805-973c2bfa2afb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.249s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.185660] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance e24c8bf2-1616-4a5a-b6ce-0c256e78b213 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.334103] env[62914]: DEBUG nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 667.691479] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance ad283bc7-b422-4836-a503-d3d9ae4fe3c1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.838201] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Acquiring lock "1d12cd15-22e9-4f96-8bd9-546196cb4138" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.838451] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Lock "1d12cd15-22e9-4f96-8bd9-546196cb4138" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.864293] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.193375] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 67554dc7-fb72-47e1-8201-f5fda035a2d3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.363869] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Acquiring lock "dcf6b4e6-3d18-4afa-be0b-a18a58a13c08" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.364786] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Lock "dcf6b4e6-3d18-4afa-be0b-a18a58a13c08" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.696364] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 821366f1-f1b0-4b78-8f7e-db8fd3bb123e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.199674] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 9edfad51-10a5-4e0c-a337-2de909c361cb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.702930] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.205976] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 8399d066-43bc-4766-97c8-89d74f3283f0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.709064] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 4c1016c4-2da8-4f4e-9497-4d680fb74535 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.212051] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 41218b72-0b79-496a-855f-57a972dabd0b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.716389] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance d9057a9e-988f-44ee-9339-ca9a1429378e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.219343] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 220578cb-17b1-4b46-8c4a-f676c04fc5e8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.723054] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance d73f5ea8-529a-44e3-8239-5aaa4534d1cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.226130] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 44339b45-5183-4882-aaec-4070adc3c3dd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.729171] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 98c19874-7c71-4e20-aa29-2ddc53df7e81 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 674.233102] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 364c5d87-f261-439a-920e-874ee0bcf83b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 674.735872] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 6201e8b7-ca3d-4eb0-8c45-67f54e66173c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.238867] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 39489032-5430-471c-8d44-2c74682ed03b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.743154] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance aa1e66d1-03fb-4340-83f1-eb79c29b6934 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.743154] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 675.743154] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 676.040530] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e149f13-5ede-482b-8701-a3af2db36ac9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.048609] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52a6774-c68d-499a-96e3-6c67f65418cf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.078195] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699bf209-9abe-41f3-97bd-6ca0a883230c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.085593] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f50a2d1-765b-495d-a8d1-c2f9a43124e5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.099024] env[62914]: DEBUG nova.compute.provider_tree [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.602027] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.107038] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 677.107312] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.995s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.107434] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.710s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.109286] env[62914]: INFO nova.compute.claims [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 678.478879] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb0c1515-1f0c-4dcd-b12c-f3b86813b7f9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.486787] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2a6060-07f5-4269-946c-e43d72a8ac91 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.516765] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b87ac79-bd1a-4424-92e0-524254ec04a9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.524851] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ffe59a-8eb7-4de0-a1be-812f8f0eb3d7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.539229] env[62914]: DEBUG nova.compute.provider_tree [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.042684] env[62914]: DEBUG nova.scheduler.client.report [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.547929] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.548516] env[62914]: DEBUG nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 679.551367] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.217s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.058647] env[62914]: DEBUG nova.compute.utils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 680.060766] env[62914]: DEBUG nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 680.060766] env[62914]: DEBUG nova.network.neutron [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 680.100361] env[62914]: DEBUG nova.policy [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab847b89cf924f228c13acc5b9d28592', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'db8a2ca3c29a49d88fb370d058f11b57', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 680.400798] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4789e538-7f3c-4403-a837-62c63bcc9efa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.408895] env[62914]: DEBUG nova.network.neutron [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Successfully created port: 8471114f-49d5-4173-90f1-d847bc66942f {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 680.411377] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed5fcd67-6760-43de-877b-c5f65661bf80 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.443316] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-680141a8-2d80-43d2-a68d-8720594c0b2d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.450693] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d92c52-5291-47ab-afb9-e8e06d8d25f0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.464422] env[62914]: DEBUG nova.compute.provider_tree [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.563438] env[62914]: DEBUG nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 680.969023] env[62914]: DEBUG nova.scheduler.client.report [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.229348] env[62914]: DEBUG nova.compute.manager [req-18b23671-d86e-41ce-8827-f5018c4c7627 req-44eda108-1b0c-4c38-b5fc-c8eaa2cfe444 service nova] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Received event network-changed-8471114f-49d5-4173-90f1-d847bc66942f {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 681.229348] env[62914]: DEBUG nova.compute.manager [req-18b23671-d86e-41ce-8827-f5018c4c7627 req-44eda108-1b0c-4c38-b5fc-c8eaa2cfe444 service nova] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Refreshing instance network info cache due to event network-changed-8471114f-49d5-4173-90f1-d847bc66942f. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 681.229348] env[62914]: DEBUG oslo_concurrency.lockutils [req-18b23671-d86e-41ce-8827-f5018c4c7627 req-44eda108-1b0c-4c38-b5fc-c8eaa2cfe444 service nova] Acquiring lock "refresh_cache-62cfca0a-e8c6-4f94-8e6f-1136cfcf3089" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.229348] env[62914]: DEBUG oslo_concurrency.lockutils [req-18b23671-d86e-41ce-8827-f5018c4c7627 req-44eda108-1b0c-4c38-b5fc-c8eaa2cfe444 service nova] Acquired lock "refresh_cache-62cfca0a-e8c6-4f94-8e6f-1136cfcf3089" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.229348] env[62914]: DEBUG nova.network.neutron [req-18b23671-d86e-41ce-8827-f5018c4c7627 req-44eda108-1b0c-4c38-b5fc-c8eaa2cfe444 service nova] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Refreshing network info cache for port 8471114f-49d5-4173-90f1-d847bc66942f {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 681.402292] env[62914]: ERROR nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8471114f-49d5-4173-90f1-d847bc66942f, please check neutron logs for more information. [ 681.402292] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 681.402292] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.402292] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 681.402292] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.402292] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 681.402292] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.402292] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 681.402292] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.402292] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 681.402292] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.402292] env[62914]: ERROR nova.compute.manager raise self.value [ 681.402292] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.402292] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 681.402292] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.402292] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 681.402748] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.402748] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 681.402748] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8471114f-49d5-4173-90f1-d847bc66942f, please check neutron logs for more information. [ 681.402748] env[62914]: ERROR nova.compute.manager [ 681.402748] env[62914]: Traceback (most recent call last): [ 681.402748] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 681.402748] env[62914]: listener.cb(fileno) [ 681.402748] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.402748] env[62914]: result = function(*args, **kwargs) [ 681.402748] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 681.402748] env[62914]: return func(*args, **kwargs) [ 681.402748] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.402748] env[62914]: raise e [ 681.402748] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.402748] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 681.402748] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.402748] env[62914]: created_port_ids = self._update_ports_for_instance( [ 681.402748] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.402748] env[62914]: with excutils.save_and_reraise_exception(): [ 681.402748] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.402748] env[62914]: self.force_reraise() [ 681.402748] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.402748] env[62914]: raise self.value [ 681.402748] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.402748] env[62914]: updated_port = self._update_port( [ 681.402748] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.402748] env[62914]: _ensure_no_port_binding_failure(port) [ 681.402748] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.402748] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 681.403488] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 8471114f-49d5-4173-90f1-d847bc66942f, please check neutron logs for more information. [ 681.403488] env[62914]: Removing descriptor: 19 [ 681.473834] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.922s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.474501] env[62914]: ERROR nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8d185603-cceb-42c2-870a-7c89e54dff99, please check neutron logs for more information. [ 681.474501] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Traceback (most recent call last): [ 681.474501] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.474501] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] self.driver.spawn(context, instance, image_meta, [ 681.474501] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 681.474501] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.474501] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.474501] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] vm_ref = self.build_virtual_machine(instance, [ 681.474501] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.474501] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.474501] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.474938] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] for vif in network_info: [ 681.474938] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.474938] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] return self._sync_wrapper(fn, *args, **kwargs) [ 681.474938] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.474938] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] self.wait() [ 681.474938] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.474938] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] self[:] = self._gt.wait() [ 681.474938] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.474938] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] return self._exit_event.wait() [ 681.474938] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 681.474938] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] current.throw(*self._exc) [ 681.474938] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.474938] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] result = function(*args, **kwargs) [ 681.475537] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 681.475537] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] return func(*args, **kwargs) [ 681.475537] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.475537] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] raise e [ 681.475537] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.475537] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] nwinfo = self.network_api.allocate_for_instance( [ 681.475537] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.475537] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] created_port_ids = self._update_ports_for_instance( [ 681.475537] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.475537] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] with excutils.save_and_reraise_exception(): [ 681.475537] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.475537] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] self.force_reraise() [ 681.475537] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.476063] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] raise self.value [ 681.476063] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.476063] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] updated_port = self._update_port( [ 681.476063] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.476063] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] _ensure_no_port_binding_failure(port) [ 681.476063] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.476063] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] raise exception.PortBindingFailed(port_id=port['id']) [ 681.476063] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] nova.exception.PortBindingFailed: Binding failed for port 8d185603-cceb-42c2-870a-7c89e54dff99, please check neutron logs for more information. [ 681.476063] env[62914]: ERROR nova.compute.manager [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] [ 681.476063] env[62914]: DEBUG nova.compute.utils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Binding failed for port 8d185603-cceb-42c2-870a-7c89e54dff99, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 681.477015] env[62914]: DEBUG oslo_concurrency.lockutils [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.099s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.477015] env[62914]: DEBUG oslo_concurrency.lockutils [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.478697] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.479s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.480095] env[62914]: INFO nova.compute.claims [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 681.482546] env[62914]: DEBUG nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Build of instance 0419073e-5340-4805-a1e0-417ccf5fdd30 was re-scheduled: Binding failed for port 8d185603-cceb-42c2-870a-7c89e54dff99, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 681.482994] env[62914]: DEBUG nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 681.483168] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Acquiring lock "refresh_cache-0419073e-5340-4805-a1e0-417ccf5fdd30" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.483312] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Acquired lock "refresh_cache-0419073e-5340-4805-a1e0-417ccf5fdd30" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.483472] env[62914]: DEBUG nova.network.neutron [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 681.506218] env[62914]: INFO nova.scheduler.client.report [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Deleted allocations for instance 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67 [ 681.576875] env[62914]: DEBUG nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 681.603955] env[62914]: DEBUG nova.virt.hardware [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 681.604233] env[62914]: DEBUG nova.virt.hardware [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 681.604399] env[62914]: DEBUG nova.virt.hardware [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 681.604579] env[62914]: DEBUG nova.virt.hardware [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 681.604721] env[62914]: DEBUG nova.virt.hardware [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 681.604862] env[62914]: DEBUG nova.virt.hardware [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 681.605072] env[62914]: DEBUG nova.virt.hardware [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 681.605346] env[62914]: DEBUG nova.virt.hardware [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 681.605402] env[62914]: DEBUG nova.virt.hardware [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 681.605577] env[62914]: DEBUG nova.virt.hardware [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 681.605754] env[62914]: DEBUG nova.virt.hardware [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 681.606656] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1885ba8-52e4-47b4-b284-eed6036db593 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.615187] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec02526-cf31-4d7b-835f-58e9bdb40e90 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.629648] env[62914]: ERROR nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8471114f-49d5-4173-90f1-d847bc66942f, please check neutron logs for more information. [ 681.629648] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Traceback (most recent call last): [ 681.629648] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 681.629648] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] yield resources [ 681.629648] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.629648] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] self.driver.spawn(context, instance, image_meta, [ 681.629648] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 681.629648] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.629648] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.629648] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] vm_ref = self.build_virtual_machine(instance, [ 681.629648] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.630029] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.630029] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.630029] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] for vif in network_info: [ 681.630029] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.630029] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] return self._sync_wrapper(fn, *args, **kwargs) [ 681.630029] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.630029] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] self.wait() [ 681.630029] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.630029] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] self[:] = self._gt.wait() [ 681.630029] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.630029] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] return self._exit_event.wait() [ 681.630029] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 681.630029] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] current.throw(*self._exc) [ 681.630451] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.630451] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] result = function(*args, **kwargs) [ 681.630451] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 681.630451] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] return func(*args, **kwargs) [ 681.630451] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.630451] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] raise e [ 681.630451] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.630451] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] nwinfo = self.network_api.allocate_for_instance( [ 681.630451] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.630451] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] created_port_ids = self._update_ports_for_instance( [ 681.630451] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.630451] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] with excutils.save_and_reraise_exception(): [ 681.630451] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.630779] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] self.force_reraise() [ 681.630779] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.630779] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] raise self.value [ 681.630779] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.630779] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] updated_port = self._update_port( [ 681.630779] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.630779] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] _ensure_no_port_binding_failure(port) [ 681.630779] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.630779] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] raise exception.PortBindingFailed(port_id=port['id']) [ 681.630779] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] nova.exception.PortBindingFailed: Binding failed for port 8471114f-49d5-4173-90f1-d847bc66942f, please check neutron logs for more information. [ 681.630779] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] [ 681.630779] env[62914]: INFO nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Terminating instance [ 681.632335] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Acquiring lock "refresh_cache-62cfca0a-e8c6-4f94-8e6f-1136cfcf3089" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.748622] env[62914]: DEBUG nova.network.neutron [req-18b23671-d86e-41ce-8827-f5018c4c7627 req-44eda108-1b0c-4c38-b5fc-c8eaa2cfe444 service nova] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.828047] env[62914]: DEBUG nova.network.neutron [req-18b23671-d86e-41ce-8827-f5018c4c7627 req-44eda108-1b0c-4c38-b5fc-c8eaa2cfe444 service nova] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.004559] env[62914]: DEBUG nova.network.neutron [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.016749] env[62914]: DEBUG oslo_concurrency.lockutils [None req-39510480-250e-486e-9d7d-99693c739d09 tempest-ServerShowV247Test-261585049 tempest-ServerShowV247Test-261585049-project-member] Lock "34b0b81e-fc98-4019-92c3-fe2ce2fdfd67" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.377s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.088769] env[62914]: DEBUG nova.network.neutron [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.330268] env[62914]: DEBUG oslo_concurrency.lockutils [req-18b23671-d86e-41ce-8827-f5018c4c7627 req-44eda108-1b0c-4c38-b5fc-c8eaa2cfe444 service nova] Releasing lock "refresh_cache-62cfca0a-e8c6-4f94-8e6f-1136cfcf3089" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.330678] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Acquired lock "refresh_cache-62cfca0a-e8c6-4f94-8e6f-1136cfcf3089" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.330861] env[62914]: DEBUG nova.network.neutron [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.593017] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Releasing lock "refresh_cache-0419073e-5340-4805-a1e0-417ccf5fdd30" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.593017] env[62914]: DEBUG nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 682.593017] env[62914]: DEBUG nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 682.593017] env[62914]: DEBUG nova.network.neutron [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 682.608250] env[62914]: DEBUG nova.network.neutron [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.814980] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1865742c-e421-4fcc-92f9-f63ab8f70ed0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.823784] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5043cb25-7fba-4a92-9b5a-1c95190fcbd9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.858040] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33cf37b3-a80c-4e49-8e46-e9f67f0e18c0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.866542] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-975beaa4-7e06-408a-9d68-b522a77265fe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.880774] env[62914]: DEBUG nova.compute.provider_tree [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.882750] env[62914]: DEBUG nova.network.neutron [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.958343] env[62914]: DEBUG nova.network.neutron [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.111885] env[62914]: DEBUG nova.network.neutron [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.252555] env[62914]: DEBUG nova.compute.manager [req-51335c15-f272-48f1-a0c7-f8ab3f568747 req-966744b7-d74e-47e6-9330-25a2df36d52b service nova] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Received event network-vif-deleted-8471114f-49d5-4173-90f1-d847bc66942f {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 683.386679] env[62914]: DEBUG nova.scheduler.client.report [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.460696] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Releasing lock "refresh_cache-62cfca0a-e8c6-4f94-8e6f-1136cfcf3089" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.461203] env[62914]: DEBUG nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 683.461392] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 683.461708] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-77d9e5ae-e352-45d3-82c8-d8cd59746196 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.472505] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789bd5d2-78fc-419d-bef0-da6d6a26eacb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.497547] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089 could not be found. [ 683.497884] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 683.498195] env[62914]: INFO nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Took 0.04 seconds to destroy the instance on the hypervisor. [ 683.498536] env[62914]: DEBUG oslo.service.loopingcall [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.498851] env[62914]: DEBUG nova.compute.manager [-] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.498998] env[62914]: DEBUG nova.network.neutron [-] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.524649] env[62914]: DEBUG nova.network.neutron [-] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.613817] env[62914]: INFO nova.compute.manager [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] [instance: 0419073e-5340-4805-a1e0-417ccf5fdd30] Took 1.02 seconds to deallocate network for instance. [ 683.892389] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.414s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.892926] env[62914]: DEBUG nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 683.895601] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.011s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.897343] env[62914]: INFO nova.compute.claims [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.027914] env[62914]: DEBUG nova.network.neutron [-] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.402471] env[62914]: DEBUG nova.compute.utils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 684.405869] env[62914]: DEBUG nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 684.405976] env[62914]: DEBUG nova.network.neutron [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 684.449196] env[62914]: DEBUG nova.policy [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '004ab0e1d0484e459b97d33e736eb93b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '76aa8b10dae94f4a8ec30728965f120a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 684.531176] env[62914]: INFO nova.compute.manager [-] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Took 1.03 seconds to deallocate network for instance. [ 684.534636] env[62914]: DEBUG nova.compute.claims [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 684.534836] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.647236] env[62914]: INFO nova.scheduler.client.report [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Deleted allocations for instance 0419073e-5340-4805-a1e0-417ccf5fdd30 [ 684.749898] env[62914]: DEBUG nova.network.neutron [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Successfully created port: 8881157e-8fc3-480e-be24-9ea26a9df050 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 684.906563] env[62914]: DEBUG nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 685.004711] env[62914]: DEBUG nova.network.neutron [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Successfully created port: 50703391-f6b0-475e-80c5-5eb0970661fb {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 685.161989] env[62914]: DEBUG oslo_concurrency.lockutils [None req-75407035-1bdf-4039-8339-2ff23d4fad97 tempest-ServerMetadataNegativeTestJSON-948317799 tempest-ServerMetadataNegativeTestJSON-948317799-project-member] Lock "0419073e-5340-4805-a1e0-417ccf5fdd30" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.562s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.288781] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6869f5f-25be-445e-923a-795ea7caeb46 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.298339] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443b3350-5c56-42ba-a33c-96772373a6a4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.329604] env[62914]: DEBUG nova.network.neutron [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Successfully created port: 1f7b047d-b81c-4377-bb08-2559a276e79a {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 685.332138] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5bf7084-e4f8-4815-b750-20b6716a3d38 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.341145] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b852f2ce-2d3a-4d1e-b8db-84c3c4454738 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.356948] env[62914]: DEBUG nova.compute.provider_tree [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.669889] env[62914]: DEBUG nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 685.860626] env[62914]: DEBUG nova.scheduler.client.report [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.921503] env[62914]: DEBUG nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 685.948565] env[62914]: DEBUG nova.virt.hardware [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 685.950019] env[62914]: DEBUG nova.virt.hardware [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 685.950019] env[62914]: DEBUG nova.virt.hardware [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 685.950019] env[62914]: DEBUG nova.virt.hardware [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 685.950019] env[62914]: DEBUG nova.virt.hardware [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 685.950019] env[62914]: DEBUG nova.virt.hardware [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 685.954022] env[62914]: DEBUG nova.virt.hardware [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 685.954022] env[62914]: DEBUG nova.virt.hardware [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 685.954022] env[62914]: DEBUG nova.virt.hardware [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 685.954022] env[62914]: DEBUG nova.virt.hardware [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 685.954022] env[62914]: DEBUG nova.virt.hardware [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 685.954415] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec07ba24-cd34-4980-846f-0f87b327ecfe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.961753] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-951da739-1095-427b-bd58-b85be3e3ad54 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.200943] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.367653] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.368263] env[62914]: DEBUG nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 686.374707] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.918s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.374707] env[62914]: INFO nova.compute.claims [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 686.380276] env[62914]: DEBUG nova.compute.manager [req-6249369b-afe0-4eb2-b4eb-a39748c622f1 req-b5dd7c69-8b17-4fec-8a3e-00bf171f9919 service nova] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Received event network-changed-8881157e-8fc3-480e-be24-9ea26a9df050 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 686.380699] env[62914]: DEBUG nova.compute.manager [req-6249369b-afe0-4eb2-b4eb-a39748c622f1 req-b5dd7c69-8b17-4fec-8a3e-00bf171f9919 service nova] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Refreshing instance network info cache due to event network-changed-8881157e-8fc3-480e-be24-9ea26a9df050. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 686.381130] env[62914]: DEBUG oslo_concurrency.lockutils [req-6249369b-afe0-4eb2-b4eb-a39748c622f1 req-b5dd7c69-8b17-4fec-8a3e-00bf171f9919 service nova] Acquiring lock "refresh_cache-294e3321-e826-4a56-bbd0-893261d65283" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.381511] env[62914]: DEBUG oslo_concurrency.lockutils [req-6249369b-afe0-4eb2-b4eb-a39748c622f1 req-b5dd7c69-8b17-4fec-8a3e-00bf171f9919 service nova] Acquired lock "refresh_cache-294e3321-e826-4a56-bbd0-893261d65283" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.381755] env[62914]: DEBUG nova.network.neutron [req-6249369b-afe0-4eb2-b4eb-a39748c622f1 req-b5dd7c69-8b17-4fec-8a3e-00bf171f9919 service nova] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Refreshing network info cache for port 8881157e-8fc3-480e-be24-9ea26a9df050 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 686.470947] env[62914]: ERROR nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8881157e-8fc3-480e-be24-9ea26a9df050, please check neutron logs for more information. [ 686.470947] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 686.470947] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.470947] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 686.470947] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.470947] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 686.470947] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.470947] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 686.470947] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.470947] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 686.470947] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.470947] env[62914]: ERROR nova.compute.manager raise self.value [ 686.470947] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.470947] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 686.470947] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.470947] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 686.471468] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.471468] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 686.471468] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8881157e-8fc3-480e-be24-9ea26a9df050, please check neutron logs for more information. [ 686.471468] env[62914]: ERROR nova.compute.manager [ 686.471468] env[62914]: Traceback (most recent call last): [ 686.471468] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 686.471468] env[62914]: listener.cb(fileno) [ 686.471468] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.471468] env[62914]: result = function(*args, **kwargs) [ 686.471468] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 686.471468] env[62914]: return func(*args, **kwargs) [ 686.471468] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.471468] env[62914]: raise e [ 686.471468] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.471468] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 686.471468] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.471468] env[62914]: created_port_ids = self._update_ports_for_instance( [ 686.471468] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.471468] env[62914]: with excutils.save_and_reraise_exception(): [ 686.471468] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.471468] env[62914]: self.force_reraise() [ 686.471468] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.471468] env[62914]: raise self.value [ 686.471468] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.471468] env[62914]: updated_port = self._update_port( [ 686.471468] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.471468] env[62914]: _ensure_no_port_binding_failure(port) [ 686.471468] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.471468] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 686.472351] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 8881157e-8fc3-480e-be24-9ea26a9df050, please check neutron logs for more information. [ 686.472351] env[62914]: Removing descriptor: 19 [ 686.472351] env[62914]: ERROR nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8881157e-8fc3-480e-be24-9ea26a9df050, please check neutron logs for more information. [ 686.472351] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] Traceback (most recent call last): [ 686.472351] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 686.472351] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] yield resources [ 686.472351] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 686.472351] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] self.driver.spawn(context, instance, image_meta, [ 686.472351] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 686.472351] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] self._vmops.spawn(context, instance, image_meta, injected_files, [ 686.472351] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 686.472351] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] vm_ref = self.build_virtual_machine(instance, [ 686.472719] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 686.472719] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] vif_infos = vmwarevif.get_vif_info(self._session, [ 686.472719] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 686.472719] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] for vif in network_info: [ 686.472719] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 686.472719] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] return self._sync_wrapper(fn, *args, **kwargs) [ 686.472719] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 686.472719] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] self.wait() [ 686.472719] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 686.472719] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] self[:] = self._gt.wait() [ 686.472719] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 686.472719] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] return self._exit_event.wait() [ 686.472719] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 686.473124] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] result = hub.switch() [ 686.473124] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 686.473124] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] return self.greenlet.switch() [ 686.473124] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.473124] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] result = function(*args, **kwargs) [ 686.473124] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 686.473124] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] return func(*args, **kwargs) [ 686.473124] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.473124] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] raise e [ 686.473124] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.473124] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] nwinfo = self.network_api.allocate_for_instance( [ 686.473124] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.473124] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] created_port_ids = self._update_ports_for_instance( [ 686.473528] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.473528] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] with excutils.save_and_reraise_exception(): [ 686.473528] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.473528] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] self.force_reraise() [ 686.473528] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.473528] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] raise self.value [ 686.473528] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.473528] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] updated_port = self._update_port( [ 686.473528] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.473528] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] _ensure_no_port_binding_failure(port) [ 686.473528] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.473528] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] raise exception.PortBindingFailed(port_id=port['id']) [ 686.473885] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] nova.exception.PortBindingFailed: Binding failed for port 8881157e-8fc3-480e-be24-9ea26a9df050, please check neutron logs for more information. [ 686.473885] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] [ 686.473885] env[62914]: INFO nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Terminating instance [ 686.474923] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Acquiring lock "refresh_cache-294e3321-e826-4a56-bbd0-893261d65283" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.885053] env[62914]: DEBUG nova.compute.utils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 686.886857] env[62914]: DEBUG nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 686.887169] env[62914]: DEBUG nova.network.neutron [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 686.942215] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Acquiring lock "cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.942441] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Lock "cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.953442] env[62914]: DEBUG nova.network.neutron [req-6249369b-afe0-4eb2-b4eb-a39748c622f1 req-b5dd7c69-8b17-4fec-8a3e-00bf171f9919 service nova] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.956918] env[62914]: DEBUG nova.policy [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '46111284e38b43528851f5fd37b8ea2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f1d308926e284f02bb0ea27b3b28902b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 687.104721] env[62914]: DEBUG nova.network.neutron [req-6249369b-afe0-4eb2-b4eb-a39748c622f1 req-b5dd7c69-8b17-4fec-8a3e-00bf171f9919 service nova] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.394616] env[62914]: DEBUG nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 687.609862] env[62914]: DEBUG oslo_concurrency.lockutils [req-6249369b-afe0-4eb2-b4eb-a39748c622f1 req-b5dd7c69-8b17-4fec-8a3e-00bf171f9919 service nova] Releasing lock "refresh_cache-294e3321-e826-4a56-bbd0-893261d65283" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.614501] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Acquired lock "refresh_cache-294e3321-e826-4a56-bbd0-893261d65283" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.614782] env[62914]: DEBUG nova.network.neutron [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 687.680630] env[62914]: DEBUG nova.network.neutron [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Successfully created port: 25685cb7-8693-45a1-8d41-212a292f6db9 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 687.815221] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45406c76-5023-4369-8cd4-aef8fe3964fd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.823581] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243cca43-118f-4aad-9b2e-ea68f7ae252b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.855614] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba33c7e5-1de7-4a79-9e6c-2743d462ba0a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.863254] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b5f213-7546-442a-b99e-59fc381e0e9e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.878770] env[62914]: DEBUG nova.compute.provider_tree [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.152046] env[62914]: DEBUG nova.network.neutron [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.258829] env[62914]: DEBUG nova.network.neutron [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.382142] env[62914]: DEBUG nova.scheduler.client.report [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.410242] env[62914]: DEBUG nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 688.419367] env[62914]: DEBUG nova.compute.manager [req-d62d2059-5a09-42c9-8918-02e8472a339b req-464841f2-846d-4b49-a220-e75c1570232e service nova] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Received event network-vif-deleted-8881157e-8fc3-480e-be24-9ea26a9df050 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 688.444119] env[62914]: DEBUG nova.virt.hardware [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 688.444119] env[62914]: DEBUG nova.virt.hardware [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 688.444119] env[62914]: DEBUG nova.virt.hardware [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 688.444367] env[62914]: DEBUG nova.virt.hardware [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 688.444367] env[62914]: DEBUG nova.virt.hardware [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 688.444367] env[62914]: DEBUG nova.virt.hardware [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 688.444367] env[62914]: DEBUG nova.virt.hardware [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 688.444367] env[62914]: DEBUG nova.virt.hardware [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 688.444541] env[62914]: DEBUG nova.virt.hardware [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 688.444720] env[62914]: DEBUG nova.virt.hardware [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 688.445031] env[62914]: DEBUG nova.virt.hardware [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 688.446477] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef238ef0-0069-49b8-9580-eecc18a9bc7f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.455707] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b97cc95-ca2b-4191-b46e-828f575e9c61 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.761700] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Releasing lock "refresh_cache-294e3321-e826-4a56-bbd0-893261d65283" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.762367] env[62914]: DEBUG nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 688.762661] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 688.763093] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4c529de8-c44a-4991-bf0b-81f00c8c5654 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.781709] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92502ec5-0359-4bd1-82e8-bad808868874 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.816986] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 294e3321-e826-4a56-bbd0-893261d65283 could not be found. [ 688.817394] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 688.817714] env[62914]: INFO nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Took 0.06 seconds to destroy the instance on the hypervisor. [ 688.818078] env[62914]: DEBUG oslo.service.loopingcall [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 688.818419] env[62914]: DEBUG nova.compute.manager [-] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 688.818556] env[62914]: DEBUG nova.network.neutron [-] [instance: 294e3321-e826-4a56-bbd0-893261d65283] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 688.855649] env[62914]: DEBUG nova.network.neutron [-] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.886901] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.886901] env[62914]: DEBUG nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 688.890515] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.976s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.890515] env[62914]: INFO nova.compute.claims [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 688.925580] env[62914]: ERROR nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 25685cb7-8693-45a1-8d41-212a292f6db9, please check neutron logs for more information. [ 688.925580] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 688.925580] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.925580] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 688.925580] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.925580] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 688.925580] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.925580] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 688.925580] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.925580] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 688.925580] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.925580] env[62914]: ERROR nova.compute.manager raise self.value [ 688.925580] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.925580] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 688.925580] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.925580] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 688.926260] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.926260] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 688.926260] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 25685cb7-8693-45a1-8d41-212a292f6db9, please check neutron logs for more information. [ 688.926260] env[62914]: ERROR nova.compute.manager [ 688.926260] env[62914]: Traceback (most recent call last): [ 688.926260] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 688.926260] env[62914]: listener.cb(fileno) [ 688.926260] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.926260] env[62914]: result = function(*args, **kwargs) [ 688.926260] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 688.926260] env[62914]: return func(*args, **kwargs) [ 688.926260] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.926260] env[62914]: raise e [ 688.926260] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.926260] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 688.926260] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.926260] env[62914]: created_port_ids = self._update_ports_for_instance( [ 688.926260] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.926260] env[62914]: with excutils.save_and_reraise_exception(): [ 688.926260] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.926260] env[62914]: self.force_reraise() [ 688.926260] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.926260] env[62914]: raise self.value [ 688.926260] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.926260] env[62914]: updated_port = self._update_port( [ 688.926260] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.926260] env[62914]: _ensure_no_port_binding_failure(port) [ 688.926260] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.926260] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 688.927120] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 25685cb7-8693-45a1-8d41-212a292f6db9, please check neutron logs for more information. [ 688.927120] env[62914]: Removing descriptor: 19 [ 688.927120] env[62914]: ERROR nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 25685cb7-8693-45a1-8d41-212a292f6db9, please check neutron logs for more information. [ 688.927120] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Traceback (most recent call last): [ 688.927120] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 688.927120] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] yield resources [ 688.927120] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 688.927120] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] self.driver.spawn(context, instance, image_meta, [ 688.927120] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 688.927120] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 688.927120] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 688.927120] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] vm_ref = self.build_virtual_machine(instance, [ 688.927656] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 688.927656] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] vif_infos = vmwarevif.get_vif_info(self._session, [ 688.927656] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 688.927656] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] for vif in network_info: [ 688.927656] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 688.927656] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] return self._sync_wrapper(fn, *args, **kwargs) [ 688.927656] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 688.927656] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] self.wait() [ 688.927656] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 688.927656] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] self[:] = self._gt.wait() [ 688.927656] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 688.927656] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] return self._exit_event.wait() [ 688.927656] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 688.927982] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] result = hub.switch() [ 688.927982] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 688.927982] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] return self.greenlet.switch() [ 688.927982] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.927982] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] result = function(*args, **kwargs) [ 688.927982] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 688.927982] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] return func(*args, **kwargs) [ 688.927982] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.927982] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] raise e [ 688.927982] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.927982] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] nwinfo = self.network_api.allocate_for_instance( [ 688.927982] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.927982] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] created_port_ids = self._update_ports_for_instance( [ 688.928853] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.928853] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] with excutils.save_and_reraise_exception(): [ 688.928853] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.928853] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] self.force_reraise() [ 688.928853] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.928853] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] raise self.value [ 688.928853] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.928853] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] updated_port = self._update_port( [ 688.928853] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.928853] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] _ensure_no_port_binding_failure(port) [ 688.928853] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.928853] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] raise exception.PortBindingFailed(port_id=port['id']) [ 688.929225] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] nova.exception.PortBindingFailed: Binding failed for port 25685cb7-8693-45a1-8d41-212a292f6db9, please check neutron logs for more information. [ 688.929225] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] [ 688.929225] env[62914]: INFO nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Terminating instance [ 688.929314] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Acquiring lock "refresh_cache-0ac803ee-6303-463f-91e8-8263c0219b0a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.929471] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Acquired lock "refresh_cache-0ac803ee-6303-463f-91e8-8263c0219b0a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.929633] env[62914]: DEBUG nova.network.neutron [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 689.395401] env[62914]: DEBUG nova.compute.utils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 689.399085] env[62914]: DEBUG nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 689.399266] env[62914]: DEBUG nova.network.neutron [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 689.455307] env[62914]: DEBUG nova.network.neutron [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.481906] env[62914]: DEBUG nova.policy [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '46111284e38b43528851f5fd37b8ea2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f1d308926e284f02bb0ea27b3b28902b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 689.549138] env[62914]: DEBUG nova.network.neutron [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.783414] env[62914]: DEBUG nova.network.neutron [-] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.884920] env[62914]: DEBUG nova.network.neutron [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Successfully created port: 26984a2f-dd05-4896-9dc6-c35d18f01310 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 689.899942] env[62914]: DEBUG nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 690.051049] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Releasing lock "refresh_cache-0ac803ee-6303-463f-91e8-8263c0219b0a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.051452] env[62914]: DEBUG nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 690.051641] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 690.053381] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-62e3c616-ef18-46df-a83e-10c8b1eb412e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.065484] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-906ed8b8-f762-4bf9-b5df-8495312994a0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.095194] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0ac803ee-6303-463f-91e8-8263c0219b0a could not be found. [ 690.095194] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 690.095194] env[62914]: INFO nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 690.095194] env[62914]: DEBUG oslo.service.loopingcall [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 690.095194] env[62914]: DEBUG nova.compute.manager [-] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 690.095194] env[62914]: DEBUG nova.network.neutron [-] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 690.113874] env[62914]: DEBUG nova.network.neutron [-] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.286906] env[62914]: INFO nova.compute.manager [-] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Took 1.47 seconds to deallocate network for instance. [ 690.296374] env[62914]: DEBUG nova.compute.claims [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 690.296580] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.309248] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab366c64-b8fb-4311-b8d9-ac8903423068 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.316873] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f93054-8a1e-43c7-b012-8246a13d64c1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.349195] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f258ea-6e3f-469c-8a75-db7da292fe42 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.356299] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b1e764f-da94-4d4a-a21c-e0fe303c7ab7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.371946] env[62914]: DEBUG nova.compute.provider_tree [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.476966] env[62914]: DEBUG nova.compute.manager [req-841d5265-2bd4-4342-80ae-24c5172906e9 req-1385551b-b291-42be-a2c3-148e3ad5b279 service nova] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Received event network-changed-25685cb7-8693-45a1-8d41-212a292f6db9 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 690.477214] env[62914]: DEBUG nova.compute.manager [req-841d5265-2bd4-4342-80ae-24c5172906e9 req-1385551b-b291-42be-a2c3-148e3ad5b279 service nova] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Refreshing instance network info cache due to event network-changed-25685cb7-8693-45a1-8d41-212a292f6db9. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 690.477395] env[62914]: DEBUG oslo_concurrency.lockutils [req-841d5265-2bd4-4342-80ae-24c5172906e9 req-1385551b-b291-42be-a2c3-148e3ad5b279 service nova] Acquiring lock "refresh_cache-0ac803ee-6303-463f-91e8-8263c0219b0a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.477556] env[62914]: DEBUG oslo_concurrency.lockutils [req-841d5265-2bd4-4342-80ae-24c5172906e9 req-1385551b-b291-42be-a2c3-148e3ad5b279 service nova] Acquired lock "refresh_cache-0ac803ee-6303-463f-91e8-8263c0219b0a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.477712] env[62914]: DEBUG nova.network.neutron [req-841d5265-2bd4-4342-80ae-24c5172906e9 req-1385551b-b291-42be-a2c3-148e3ad5b279 service nova] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Refreshing network info cache for port 25685cb7-8693-45a1-8d41-212a292f6db9 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 690.616429] env[62914]: DEBUG nova.network.neutron [-] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.877017] env[62914]: DEBUG nova.scheduler.client.report [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.913636] env[62914]: DEBUG nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 690.949409] env[62914]: DEBUG nova.virt.hardware [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 690.949643] env[62914]: DEBUG nova.virt.hardware [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 690.949788] env[62914]: DEBUG nova.virt.hardware [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 690.950047] env[62914]: DEBUG nova.virt.hardware [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 690.950341] env[62914]: DEBUG nova.virt.hardware [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 690.950529] env[62914]: DEBUG nova.virt.hardware [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 690.950737] env[62914]: DEBUG nova.virt.hardware [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 690.950895] env[62914]: DEBUG nova.virt.hardware [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 690.951270] env[62914]: DEBUG nova.virt.hardware [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 690.951464] env[62914]: DEBUG nova.virt.hardware [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 690.952274] env[62914]: DEBUG nova.virt.hardware [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 690.952520] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbaecc10-2d11-4914-a860-16492402a95e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.961551] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde7262d-89ae-4a07-b6db-1f2c70607e04 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.004819] env[62914]: DEBUG nova.network.neutron [req-841d5265-2bd4-4342-80ae-24c5172906e9 req-1385551b-b291-42be-a2c3-148e3ad5b279 service nova] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.062320] env[62914]: DEBUG nova.network.neutron [req-841d5265-2bd4-4342-80ae-24c5172906e9 req-1385551b-b291-42be-a2c3-148e3ad5b279 service nova] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.118492] env[62914]: INFO nova.compute.manager [-] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Took 1.02 seconds to deallocate network for instance. [ 691.120919] env[62914]: DEBUG nova.compute.claims [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 691.121113] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.383060] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.383592] env[62914]: DEBUG nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 691.386659] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.532s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.388090] env[62914]: INFO nova.compute.claims [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 691.555113] env[62914]: DEBUG nova.compute.manager [req-2145a3da-d8d3-43cc-89d9-c4d294fe8c40 req-93efc0ca-c8d8-4090-9c28-b36472aca011 service nova] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Received event network-changed-26984a2f-dd05-4896-9dc6-c35d18f01310 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 691.555367] env[62914]: DEBUG nova.compute.manager [req-2145a3da-d8d3-43cc-89d9-c4d294fe8c40 req-93efc0ca-c8d8-4090-9c28-b36472aca011 service nova] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Refreshing instance network info cache due to event network-changed-26984a2f-dd05-4896-9dc6-c35d18f01310. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 691.555520] env[62914]: DEBUG oslo_concurrency.lockutils [req-2145a3da-d8d3-43cc-89d9-c4d294fe8c40 req-93efc0ca-c8d8-4090-9c28-b36472aca011 service nova] Acquiring lock "refresh_cache-3ce62ec7-6274-441f-989b-6f27a64b0c18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.555658] env[62914]: DEBUG oslo_concurrency.lockutils [req-2145a3da-d8d3-43cc-89d9-c4d294fe8c40 req-93efc0ca-c8d8-4090-9c28-b36472aca011 service nova] Acquired lock "refresh_cache-3ce62ec7-6274-441f-989b-6f27a64b0c18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.555875] env[62914]: DEBUG nova.network.neutron [req-2145a3da-d8d3-43cc-89d9-c4d294fe8c40 req-93efc0ca-c8d8-4090-9c28-b36472aca011 service nova] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Refreshing network info cache for port 26984a2f-dd05-4896-9dc6-c35d18f01310 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 691.565849] env[62914]: DEBUG oslo_concurrency.lockutils [req-841d5265-2bd4-4342-80ae-24c5172906e9 req-1385551b-b291-42be-a2c3-148e3ad5b279 service nova] Releasing lock "refresh_cache-0ac803ee-6303-463f-91e8-8263c0219b0a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.566064] env[62914]: DEBUG nova.compute.manager [req-841d5265-2bd4-4342-80ae-24c5172906e9 req-1385551b-b291-42be-a2c3-148e3ad5b279 service nova] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Received event network-vif-deleted-25685cb7-8693-45a1-8d41-212a292f6db9 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 691.730102] env[62914]: ERROR nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 26984a2f-dd05-4896-9dc6-c35d18f01310, please check neutron logs for more information. [ 691.730102] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 691.730102] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.730102] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 691.730102] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.730102] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 691.730102] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.730102] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 691.730102] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.730102] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 691.730102] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.730102] env[62914]: ERROR nova.compute.manager raise self.value [ 691.730102] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.730102] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 691.730102] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.730102] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 691.730977] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.730977] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 691.730977] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 26984a2f-dd05-4896-9dc6-c35d18f01310, please check neutron logs for more information. [ 691.730977] env[62914]: ERROR nova.compute.manager [ 691.730977] env[62914]: Traceback (most recent call last): [ 691.730977] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 691.730977] env[62914]: listener.cb(fileno) [ 691.730977] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.730977] env[62914]: result = function(*args, **kwargs) [ 691.730977] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 691.730977] env[62914]: return func(*args, **kwargs) [ 691.730977] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.730977] env[62914]: raise e [ 691.730977] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.730977] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 691.730977] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.730977] env[62914]: created_port_ids = self._update_ports_for_instance( [ 691.730977] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.730977] env[62914]: with excutils.save_and_reraise_exception(): [ 691.730977] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.730977] env[62914]: self.force_reraise() [ 691.730977] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.730977] env[62914]: raise self.value [ 691.730977] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.730977] env[62914]: updated_port = self._update_port( [ 691.730977] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.730977] env[62914]: _ensure_no_port_binding_failure(port) [ 691.730977] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.730977] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 691.732137] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 26984a2f-dd05-4896-9dc6-c35d18f01310, please check neutron logs for more information. [ 691.732137] env[62914]: Removing descriptor: 15 [ 691.732137] env[62914]: ERROR nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 26984a2f-dd05-4896-9dc6-c35d18f01310, please check neutron logs for more information. [ 691.732137] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Traceback (most recent call last): [ 691.732137] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 691.732137] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] yield resources [ 691.732137] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 691.732137] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] self.driver.spawn(context, instance, image_meta, [ 691.732137] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 691.732137] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] self._vmops.spawn(context, instance, image_meta, injected_files, [ 691.732137] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 691.732137] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] vm_ref = self.build_virtual_machine(instance, [ 691.732677] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 691.732677] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] vif_infos = vmwarevif.get_vif_info(self._session, [ 691.732677] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 691.732677] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] for vif in network_info: [ 691.732677] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 691.732677] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] return self._sync_wrapper(fn, *args, **kwargs) [ 691.732677] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 691.732677] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] self.wait() [ 691.732677] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 691.732677] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] self[:] = self._gt.wait() [ 691.732677] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 691.732677] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] return self._exit_event.wait() [ 691.732677] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 691.733151] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] result = hub.switch() [ 691.733151] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 691.733151] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] return self.greenlet.switch() [ 691.733151] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.733151] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] result = function(*args, **kwargs) [ 691.733151] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 691.733151] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] return func(*args, **kwargs) [ 691.733151] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.733151] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] raise e [ 691.733151] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.733151] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] nwinfo = self.network_api.allocate_for_instance( [ 691.733151] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.733151] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] created_port_ids = self._update_ports_for_instance( [ 691.733509] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.733509] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] with excutils.save_and_reraise_exception(): [ 691.733509] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.733509] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] self.force_reraise() [ 691.733509] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.733509] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] raise self.value [ 691.733509] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.733509] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] updated_port = self._update_port( [ 691.733509] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.733509] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] _ensure_no_port_binding_failure(port) [ 691.733509] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.733509] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] raise exception.PortBindingFailed(port_id=port['id']) [ 691.733836] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] nova.exception.PortBindingFailed: Binding failed for port 26984a2f-dd05-4896-9dc6-c35d18f01310, please check neutron logs for more information. [ 691.733836] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] [ 691.733836] env[62914]: INFO nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Terminating instance [ 691.737110] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Acquiring lock "refresh_cache-3ce62ec7-6274-441f-989b-6f27a64b0c18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.892962] env[62914]: DEBUG nova.compute.utils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 691.896671] env[62914]: DEBUG nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 691.896879] env[62914]: DEBUG nova.network.neutron [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 691.944014] env[62914]: DEBUG nova.policy [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1fb70900c272422583fd50fc271c65cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd3ee6294864e4d6fa9abcf3199aa1083', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 692.076214] env[62914]: DEBUG nova.network.neutron [req-2145a3da-d8d3-43cc-89d9-c4d294fe8c40 req-93efc0ca-c8d8-4090-9c28-b36472aca011 service nova] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.142771] env[62914]: DEBUG nova.network.neutron [req-2145a3da-d8d3-43cc-89d9-c4d294fe8c40 req-93efc0ca-c8d8-4090-9c28-b36472aca011 service nova] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.237403] env[62914]: DEBUG nova.network.neutron [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Successfully created port: a6637332-9313-492b-bdde-1386cdbb1191 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 692.397900] env[62914]: DEBUG nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 692.571544] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Acquiring lock "cd5a266b-617d-47c5-adb9-d161bd68f672" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.571795] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Lock "cd5a266b-617d-47c5-adb9-d161bd68f672" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.648676] env[62914]: DEBUG oslo_concurrency.lockutils [req-2145a3da-d8d3-43cc-89d9-c4d294fe8c40 req-93efc0ca-c8d8-4090-9c28-b36472aca011 service nova] Releasing lock "refresh_cache-3ce62ec7-6274-441f-989b-6f27a64b0c18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.649763] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Acquired lock "refresh_cache-3ce62ec7-6274-441f-989b-6f27a64b0c18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.651958] env[62914]: DEBUG nova.network.neutron [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 692.841315] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f1fc3d-975b-45d5-83d1-9be2247b1275 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.850083] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8690e4b2-f13f-4bd8-977a-52810c7c0d67 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.883985] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce2d8c40-aaa1-4b03-811b-21c6a8479366 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.889914] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c69943-c7c5-4d60-8377-88f8da9c7763 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.903109] env[62914]: DEBUG nova.compute.provider_tree [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.172329] env[62914]: DEBUG nova.network.neutron [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.227763] env[62914]: DEBUG nova.network.neutron [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.305168] env[62914]: ERROR nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a6637332-9313-492b-bdde-1386cdbb1191, please check neutron logs for more information. [ 693.305168] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 693.305168] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.305168] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 693.305168] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.305168] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 693.305168] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.305168] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 693.305168] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.305168] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 693.305168] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.305168] env[62914]: ERROR nova.compute.manager raise self.value [ 693.305168] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.305168] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 693.305168] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.305168] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 693.305702] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.305702] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 693.305702] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a6637332-9313-492b-bdde-1386cdbb1191, please check neutron logs for more information. [ 693.305702] env[62914]: ERROR nova.compute.manager [ 693.305702] env[62914]: Traceback (most recent call last): [ 693.305702] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 693.305702] env[62914]: listener.cb(fileno) [ 693.305702] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.305702] env[62914]: result = function(*args, **kwargs) [ 693.305702] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 693.305702] env[62914]: return func(*args, **kwargs) [ 693.305702] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.305702] env[62914]: raise e [ 693.305702] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.305702] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 693.305702] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.305702] env[62914]: created_port_ids = self._update_ports_for_instance( [ 693.305702] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.305702] env[62914]: with excutils.save_and_reraise_exception(): [ 693.305702] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.305702] env[62914]: self.force_reraise() [ 693.305702] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.305702] env[62914]: raise self.value [ 693.305702] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.305702] env[62914]: updated_port = self._update_port( [ 693.305702] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.305702] env[62914]: _ensure_no_port_binding_failure(port) [ 693.305702] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.305702] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 693.306600] env[62914]: nova.exception.PortBindingFailed: Binding failed for port a6637332-9313-492b-bdde-1386cdbb1191, please check neutron logs for more information. [ 693.306600] env[62914]: Removing descriptor: 15 [ 693.409160] env[62914]: DEBUG nova.scheduler.client.report [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 693.412907] env[62914]: DEBUG nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 693.439375] env[62914]: DEBUG nova.virt.hardware [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 693.439620] env[62914]: DEBUG nova.virt.hardware [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 693.439772] env[62914]: DEBUG nova.virt.hardware [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 693.439948] env[62914]: DEBUG nova.virt.hardware [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 693.440102] env[62914]: DEBUG nova.virt.hardware [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 693.440248] env[62914]: DEBUG nova.virt.hardware [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 693.440453] env[62914]: DEBUG nova.virt.hardware [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 693.440606] env[62914]: DEBUG nova.virt.hardware [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 693.440765] env[62914]: DEBUG nova.virt.hardware [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 693.440922] env[62914]: DEBUG nova.virt.hardware [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 693.441164] env[62914]: DEBUG nova.virt.hardware [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 693.442351] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0092ad27-8560-4960-9741-7bffa05016ae {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.452016] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-700be5b5-710c-49b5-87ba-8ded6b04e06f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.465644] env[62914]: ERROR nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a6637332-9313-492b-bdde-1386cdbb1191, please check neutron logs for more information. [ 693.465644] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Traceback (most recent call last): [ 693.465644] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 693.465644] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] yield resources [ 693.465644] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 693.465644] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] self.driver.spawn(context, instance, image_meta, [ 693.465644] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 693.465644] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] self._vmops.spawn(context, instance, image_meta, injected_files, [ 693.465644] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 693.465644] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] vm_ref = self.build_virtual_machine(instance, [ 693.465644] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 693.465963] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] vif_infos = vmwarevif.get_vif_info(self._session, [ 693.465963] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 693.465963] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] for vif in network_info: [ 693.465963] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 693.465963] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] return self._sync_wrapper(fn, *args, **kwargs) [ 693.465963] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 693.465963] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] self.wait() [ 693.465963] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 693.465963] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] self[:] = self._gt.wait() [ 693.465963] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 693.465963] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] return self._exit_event.wait() [ 693.465963] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 693.465963] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] current.throw(*self._exc) [ 693.466334] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.466334] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] result = function(*args, **kwargs) [ 693.466334] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 693.466334] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] return func(*args, **kwargs) [ 693.466334] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.466334] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] raise e [ 693.466334] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.466334] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] nwinfo = self.network_api.allocate_for_instance( [ 693.466334] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.466334] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] created_port_ids = self._update_ports_for_instance( [ 693.466334] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.466334] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] with excutils.save_and_reraise_exception(): [ 693.466334] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.466765] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] self.force_reraise() [ 693.466765] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.466765] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] raise self.value [ 693.466765] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.466765] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] updated_port = self._update_port( [ 693.466765] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.466765] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] _ensure_no_port_binding_failure(port) [ 693.466765] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.466765] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] raise exception.PortBindingFailed(port_id=port['id']) [ 693.466765] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] nova.exception.PortBindingFailed: Binding failed for port a6637332-9313-492b-bdde-1386cdbb1191, please check neutron logs for more information. [ 693.466765] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] [ 693.466765] env[62914]: INFO nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Terminating instance [ 693.468023] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Acquiring lock "refresh_cache-e24c8bf2-1616-4a5a-b6ce-0c256e78b213" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.468191] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Acquired lock "refresh_cache-e24c8bf2-1616-4a5a-b6ce-0c256e78b213" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.468350] env[62914]: DEBUG nova.network.neutron [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 693.583320] env[62914]: DEBUG nova.compute.manager [req-7ef055e8-33f0-47bc-b174-ff7fa0a85eaa req-893fa113-a6f7-4fdf-b88a-847795e640bd service nova] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Received event network-vif-deleted-26984a2f-dd05-4896-9dc6-c35d18f01310 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 693.583594] env[62914]: DEBUG nova.compute.manager [req-7ef055e8-33f0-47bc-b174-ff7fa0a85eaa req-893fa113-a6f7-4fdf-b88a-847795e640bd service nova] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Received event network-changed-a6637332-9313-492b-bdde-1386cdbb1191 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 693.583969] env[62914]: DEBUG nova.compute.manager [req-7ef055e8-33f0-47bc-b174-ff7fa0a85eaa req-893fa113-a6f7-4fdf-b88a-847795e640bd service nova] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Refreshing instance network info cache due to event network-changed-a6637332-9313-492b-bdde-1386cdbb1191. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 693.584062] env[62914]: DEBUG oslo_concurrency.lockutils [req-7ef055e8-33f0-47bc-b174-ff7fa0a85eaa req-893fa113-a6f7-4fdf-b88a-847795e640bd service nova] Acquiring lock "refresh_cache-e24c8bf2-1616-4a5a-b6ce-0c256e78b213" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.730841] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Releasing lock "refresh_cache-3ce62ec7-6274-441f-989b-6f27a64b0c18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.731359] env[62914]: DEBUG nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 693.731599] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 693.731944] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4e39e2e-eb66-4493-9232-2004aaac0124 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.741307] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b7496c-eee0-4146-942e-e36177c0ad7b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.762763] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3ce62ec7-6274-441f-989b-6f27a64b0c18 could not be found. [ 693.762968] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 693.763158] env[62914]: INFO nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Took 0.03 seconds to destroy the instance on the hypervisor. [ 693.763394] env[62914]: DEBUG oslo.service.loopingcall [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.763606] env[62914]: DEBUG nova.compute.manager [-] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 693.763700] env[62914]: DEBUG nova.network.neutron [-] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 693.785702] env[62914]: DEBUG nova.network.neutron [-] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.916387] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.530s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.917009] env[62914]: DEBUG nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 693.920039] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.905s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.983883] env[62914]: DEBUG nova.network.neutron [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.058432] env[62914]: DEBUG nova.network.neutron [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.288273] env[62914]: DEBUG nova.network.neutron [-] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.424768] env[62914]: DEBUG nova.compute.utils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 694.426158] env[62914]: DEBUG nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 694.426325] env[62914]: DEBUG nova.network.neutron [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 694.477426] env[62914]: DEBUG nova.policy [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd939241d75294f38ba67fc1bb9cb59cc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '230784fa3f0f4b9c980e2cef5cb39ef8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 694.560317] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Releasing lock "refresh_cache-e24c8bf2-1616-4a5a-b6ce-0c256e78b213" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.560732] env[62914]: DEBUG nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 694.560924] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 694.561245] env[62914]: DEBUG oslo_concurrency.lockutils [req-7ef055e8-33f0-47bc-b174-ff7fa0a85eaa req-893fa113-a6f7-4fdf-b88a-847795e640bd service nova] Acquired lock "refresh_cache-e24c8bf2-1616-4a5a-b6ce-0c256e78b213" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.561414] env[62914]: DEBUG nova.network.neutron [req-7ef055e8-33f0-47bc-b174-ff7fa0a85eaa req-893fa113-a6f7-4fdf-b88a-847795e640bd service nova] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Refreshing network info cache for port a6637332-9313-492b-bdde-1386cdbb1191 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 694.562463] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c46a6164-15b9-4bff-b17f-fc438ce3996d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.575726] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43bcae00-42ae-4d04-af2e-0ddd3eae599b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.599460] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e24c8bf2-1616-4a5a-b6ce-0c256e78b213 could not be found. [ 694.599704] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 694.599842] env[62914]: INFO nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Took 0.04 seconds to destroy the instance on the hypervisor. [ 694.600104] env[62914]: DEBUG oslo.service.loopingcall [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 694.602408] env[62914]: DEBUG nova.compute.manager [-] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.602508] env[62914]: DEBUG nova.network.neutron [-] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 694.617805] env[62914]: DEBUG nova.network.neutron [-] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.768076] env[62914]: DEBUG nova.network.neutron [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Successfully created port: 48a08124-eb4f-427e-92bb-2c75fa5649a3 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 694.796528] env[62914]: INFO nova.compute.manager [-] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Took 1.03 seconds to deallocate network for instance. [ 694.799636] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e273e2b9-97b8-4ec9-8389-0060c2775b98 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.802419] env[62914]: DEBUG nova.compute.claims [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 694.802592] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.807880] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bacaefa2-c8a8-4565-ba28-cfd8eff7cbed {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.837773] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be84f77e-2de9-4f18-a58d-cef1cd2e8e77 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.845635] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a38db1f9-e854-4f87-9085-fa14d100c86f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.860007] env[62914]: DEBUG nova.compute.provider_tree [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.931866] env[62914]: DEBUG nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 695.083814] env[62914]: DEBUG nova.network.neutron [req-7ef055e8-33f0-47bc-b174-ff7fa0a85eaa req-893fa113-a6f7-4fdf-b88a-847795e640bd service nova] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.119612] env[62914]: DEBUG nova.network.neutron [-] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.165551] env[62914]: DEBUG nova.network.neutron [req-7ef055e8-33f0-47bc-b174-ff7fa0a85eaa req-893fa113-a6f7-4fdf-b88a-847795e640bd service nova] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.365574] env[62914]: DEBUG nova.scheduler.client.report [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 695.621394] env[62914]: DEBUG nova.compute.manager [req-25a7f0ae-0b30-4398-89aa-05b492fbc748 req-701b1033-47aa-4c47-9eaf-a1318132164f service nova] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Received event network-changed-48a08124-eb4f-427e-92bb-2c75fa5649a3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 695.621642] env[62914]: DEBUG nova.compute.manager [req-25a7f0ae-0b30-4398-89aa-05b492fbc748 req-701b1033-47aa-4c47-9eaf-a1318132164f service nova] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Refreshing instance network info cache due to event network-changed-48a08124-eb4f-427e-92bb-2c75fa5649a3. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 695.621884] env[62914]: DEBUG oslo_concurrency.lockutils [req-25a7f0ae-0b30-4398-89aa-05b492fbc748 req-701b1033-47aa-4c47-9eaf-a1318132164f service nova] Acquiring lock "refresh_cache-ad283bc7-b422-4836-a503-d3d9ae4fe3c1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.622076] env[62914]: DEBUG oslo_concurrency.lockutils [req-25a7f0ae-0b30-4398-89aa-05b492fbc748 req-701b1033-47aa-4c47-9eaf-a1318132164f service nova] Acquired lock "refresh_cache-ad283bc7-b422-4836-a503-d3d9ae4fe3c1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.622275] env[62914]: DEBUG nova.network.neutron [req-25a7f0ae-0b30-4398-89aa-05b492fbc748 req-701b1033-47aa-4c47-9eaf-a1318132164f service nova] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Refreshing network info cache for port 48a08124-eb4f-427e-92bb-2c75fa5649a3 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 695.625899] env[62914]: INFO nova.compute.manager [-] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Took 1.02 seconds to deallocate network for instance. [ 695.627881] env[62914]: DEBUG nova.compute.claims [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 695.627881] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.668719] env[62914]: DEBUG oslo_concurrency.lockutils [req-7ef055e8-33f0-47bc-b174-ff7fa0a85eaa req-893fa113-a6f7-4fdf-b88a-847795e640bd service nova] Releasing lock "refresh_cache-e24c8bf2-1616-4a5a-b6ce-0c256e78b213" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.668719] env[62914]: DEBUG nova.compute.manager [req-7ef055e8-33f0-47bc-b174-ff7fa0a85eaa req-893fa113-a6f7-4fdf-b88a-847795e640bd service nova] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Received event network-vif-deleted-a6637332-9313-492b-bdde-1386cdbb1191 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 695.740407] env[62914]: ERROR nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 48a08124-eb4f-427e-92bb-2c75fa5649a3, please check neutron logs for more information. [ 695.740407] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 695.740407] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.740407] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 695.740407] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.740407] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 695.740407] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.740407] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 695.740407] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.740407] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 695.740407] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.740407] env[62914]: ERROR nova.compute.manager raise self.value [ 695.740407] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.740407] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 695.740407] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.740407] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 695.740888] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.740888] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 695.740888] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 48a08124-eb4f-427e-92bb-2c75fa5649a3, please check neutron logs for more information. [ 695.740888] env[62914]: ERROR nova.compute.manager [ 695.740888] env[62914]: Traceback (most recent call last): [ 695.740888] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 695.740888] env[62914]: listener.cb(fileno) [ 695.740888] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.740888] env[62914]: result = function(*args, **kwargs) [ 695.740888] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 695.740888] env[62914]: return func(*args, **kwargs) [ 695.740888] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.740888] env[62914]: raise e [ 695.740888] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.740888] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 695.740888] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.740888] env[62914]: created_port_ids = self._update_ports_for_instance( [ 695.740888] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.740888] env[62914]: with excutils.save_and_reraise_exception(): [ 695.740888] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.740888] env[62914]: self.force_reraise() [ 695.740888] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.740888] env[62914]: raise self.value [ 695.740888] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.740888] env[62914]: updated_port = self._update_port( [ 695.740888] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.740888] env[62914]: _ensure_no_port_binding_failure(port) [ 695.740888] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.740888] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 695.741959] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 48a08124-eb4f-427e-92bb-2c75fa5649a3, please check neutron logs for more information. [ 695.741959] env[62914]: Removing descriptor: 15 [ 695.872773] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.953s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.873408] env[62914]: ERROR nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7fef89a0-fac4-420f-8c07-a91a671d5de9, please check neutron logs for more information. [ 695.873408] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Traceback (most recent call last): [ 695.873408] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 695.873408] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] self.driver.spawn(context, instance, image_meta, [ 695.873408] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 695.873408] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.873408] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.873408] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] vm_ref = self.build_virtual_machine(instance, [ 695.873408] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.873408] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.873408] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.873753] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] for vif in network_info: [ 695.873753] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.873753] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] return self._sync_wrapper(fn, *args, **kwargs) [ 695.873753] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.873753] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] self.wait() [ 695.873753] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.873753] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] self[:] = self._gt.wait() [ 695.873753] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.873753] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] return self._exit_event.wait() [ 695.873753] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 695.873753] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] result = hub.switch() [ 695.873753] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 695.873753] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] return self.greenlet.switch() [ 695.874160] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.874160] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] result = function(*args, **kwargs) [ 695.874160] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 695.874160] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] return func(*args, **kwargs) [ 695.874160] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.874160] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] raise e [ 695.874160] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.874160] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] nwinfo = self.network_api.allocate_for_instance( [ 695.874160] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.874160] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] created_port_ids = self._update_ports_for_instance( [ 695.874160] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.874160] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] with excutils.save_and_reraise_exception(): [ 695.874160] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.874541] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] self.force_reraise() [ 695.874541] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.874541] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] raise self.value [ 695.874541] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.874541] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] updated_port = self._update_port( [ 695.874541] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.874541] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] _ensure_no_port_binding_failure(port) [ 695.874541] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.874541] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] raise exception.PortBindingFailed(port_id=port['id']) [ 695.874541] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] nova.exception.PortBindingFailed: Binding failed for port 7fef89a0-fac4-420f-8c07-a91a671d5de9, please check neutron logs for more information. [ 695.874541] env[62914]: ERROR nova.compute.manager [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] [ 695.874877] env[62914]: DEBUG nova.compute.utils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Binding failed for port 7fef89a0-fac4-420f-8c07-a91a671d5de9, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 695.875303] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.246s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.878255] env[62914]: DEBUG nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Build of instance c10d840e-ce05-4af3-ba2d-6214eefd8783 was re-scheduled: Binding failed for port 7fef89a0-fac4-420f-8c07-a91a671d5de9, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 695.878662] env[62914]: DEBUG nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 695.878884] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Acquiring lock "refresh_cache-c10d840e-ce05-4af3-ba2d-6214eefd8783" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.879044] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Acquired lock "refresh_cache-c10d840e-ce05-4af3-ba2d-6214eefd8783" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.879201] env[62914]: DEBUG nova.network.neutron [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.942023] env[62914]: DEBUG nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 695.981392] env[62914]: DEBUG nova.virt.hardware [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:09:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='568721fd-c1c8-43d2-bc51-18e26e21676b',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1657649571',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 695.981649] env[62914]: DEBUG nova.virt.hardware [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 695.981802] env[62914]: DEBUG nova.virt.hardware [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 695.981980] env[62914]: DEBUG nova.virt.hardware [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 695.982140] env[62914]: DEBUG nova.virt.hardware [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 695.982285] env[62914]: DEBUG nova.virt.hardware [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 695.982497] env[62914]: DEBUG nova.virt.hardware [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 695.982657] env[62914]: DEBUG nova.virt.hardware [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 695.982815] env[62914]: DEBUG nova.virt.hardware [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 695.982979] env[62914]: DEBUG nova.virt.hardware [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 695.983153] env[62914]: DEBUG nova.virt.hardware [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 695.984013] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3caadb61-c7a9-457f-af9a-285706cf0138 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.992189] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-455ed41b-7393-4abb-883e-035a0187bf81 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.005504] env[62914]: ERROR nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 48a08124-eb4f-427e-92bb-2c75fa5649a3, please check neutron logs for more information. [ 696.005504] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Traceback (most recent call last): [ 696.005504] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 696.005504] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] yield resources [ 696.005504] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 696.005504] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] self.driver.spawn(context, instance, image_meta, [ 696.005504] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 696.005504] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.005504] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.005504] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] vm_ref = self.build_virtual_machine(instance, [ 696.005504] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.005900] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.005900] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.005900] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] for vif in network_info: [ 696.005900] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.005900] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] return self._sync_wrapper(fn, *args, **kwargs) [ 696.005900] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.005900] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] self.wait() [ 696.005900] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.005900] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] self[:] = self._gt.wait() [ 696.005900] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.005900] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] return self._exit_event.wait() [ 696.005900] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 696.005900] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] current.throw(*self._exc) [ 696.006317] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.006317] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] result = function(*args, **kwargs) [ 696.006317] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 696.006317] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] return func(*args, **kwargs) [ 696.006317] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.006317] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] raise e [ 696.006317] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.006317] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] nwinfo = self.network_api.allocate_for_instance( [ 696.006317] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 696.006317] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] created_port_ids = self._update_ports_for_instance( [ 696.006317] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 696.006317] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] with excutils.save_and_reraise_exception(): [ 696.006317] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.006739] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] self.force_reraise() [ 696.006739] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.006739] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] raise self.value [ 696.006739] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 696.006739] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] updated_port = self._update_port( [ 696.006739] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.006739] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] _ensure_no_port_binding_failure(port) [ 696.006739] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.006739] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] raise exception.PortBindingFailed(port_id=port['id']) [ 696.006739] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] nova.exception.PortBindingFailed: Binding failed for port 48a08124-eb4f-427e-92bb-2c75fa5649a3, please check neutron logs for more information. [ 696.006739] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] [ 696.006739] env[62914]: INFO nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Terminating instance [ 696.007817] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Acquiring lock "refresh_cache-ad283bc7-b422-4836-a503-d3d9ae4fe3c1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.140447] env[62914]: DEBUG nova.network.neutron [req-25a7f0ae-0b30-4398-89aa-05b492fbc748 req-701b1033-47aa-4c47-9eaf-a1318132164f service nova] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.189316] env[62914]: DEBUG nova.network.neutron [req-25a7f0ae-0b30-4398-89aa-05b492fbc748 req-701b1033-47aa-4c47-9eaf-a1318132164f service nova] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.398652] env[62914]: DEBUG nova.network.neutron [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.480145] env[62914]: DEBUG nova.network.neutron [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.688167] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84433804-dac7-443b-8240-f9376dae205d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.692328] env[62914]: DEBUG oslo_concurrency.lockutils [req-25a7f0ae-0b30-4398-89aa-05b492fbc748 req-701b1033-47aa-4c47-9eaf-a1318132164f service nova] Releasing lock "refresh_cache-ad283bc7-b422-4836-a503-d3d9ae4fe3c1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.692693] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Acquired lock "refresh_cache-ad283bc7-b422-4836-a503-d3d9ae4fe3c1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.692875] env[62914]: DEBUG nova.network.neutron [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 696.696790] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa18f34-ea79-4785-b629-bc5b6e63ff3e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.726995] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c2f889-7af6-4e09-83c3-fae1699beb32 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.733942] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd342f8d-2d83-4eec-b2df-89405c43fe6b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.746799] env[62914]: DEBUG nova.compute.provider_tree [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.983996] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Releasing lock "refresh_cache-c10d840e-ce05-4af3-ba2d-6214eefd8783" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.984235] env[62914]: DEBUG nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 696.984421] env[62914]: DEBUG nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.984584] env[62914]: DEBUG nova.network.neutron [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 697.000467] env[62914]: DEBUG nova.network.neutron [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.210799] env[62914]: DEBUG nova.network.neutron [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.249730] env[62914]: DEBUG nova.scheduler.client.report [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.265295] env[62914]: DEBUG nova.network.neutron [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.503220] env[62914]: DEBUG nova.network.neutron [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.646103] env[62914]: DEBUG nova.compute.manager [req-ed445b79-cec7-4e84-8d23-d5250389c554 req-1e8cdfd8-f410-48cf-a25a-5e211014750b service nova] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Received event network-vif-deleted-48a08124-eb4f-427e-92bb-2c75fa5649a3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 697.754458] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.879s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.755124] env[62914]: ERROR nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 83e791da-9e3a-4c0f-9454-6662da4e8798, please check neutron logs for more information. [ 697.755124] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Traceback (most recent call last): [ 697.755124] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 697.755124] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] self.driver.spawn(context, instance, image_meta, [ 697.755124] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 697.755124] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] self._vmops.spawn(context, instance, image_meta, injected_files, [ 697.755124] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 697.755124] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] vm_ref = self.build_virtual_machine(instance, [ 697.755124] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 697.755124] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] vif_infos = vmwarevif.get_vif_info(self._session, [ 697.755124] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 697.755563] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] for vif in network_info: [ 697.755563] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 697.755563] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] return self._sync_wrapper(fn, *args, **kwargs) [ 697.755563] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 697.755563] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] self.wait() [ 697.755563] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 697.755563] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] self[:] = self._gt.wait() [ 697.755563] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 697.755563] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] return self._exit_event.wait() [ 697.755563] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 697.755563] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] result = hub.switch() [ 697.755563] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 697.755563] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] return self.greenlet.switch() [ 697.755944] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.755944] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] result = function(*args, **kwargs) [ 697.755944] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 697.755944] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] return func(*args, **kwargs) [ 697.755944] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.755944] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] raise e [ 697.755944] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.755944] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] nwinfo = self.network_api.allocate_for_instance( [ 697.755944] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.755944] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] created_port_ids = self._update_ports_for_instance( [ 697.755944] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.755944] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] with excutils.save_and_reraise_exception(): [ 697.755944] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.756353] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] self.force_reraise() [ 697.756353] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.756353] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] raise self.value [ 697.756353] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.756353] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] updated_port = self._update_port( [ 697.756353] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.756353] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] _ensure_no_port_binding_failure(port) [ 697.756353] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.756353] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] raise exception.PortBindingFailed(port_id=port['id']) [ 697.756353] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] nova.exception.PortBindingFailed: Binding failed for port 83e791da-9e3a-4c0f-9454-6662da4e8798, please check neutron logs for more information. [ 697.756353] env[62914]: ERROR nova.compute.manager [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] [ 697.756731] env[62914]: DEBUG nova.compute.utils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Binding failed for port 83e791da-9e3a-4c0f-9454-6662da4e8798, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 697.757491] env[62914]: DEBUG nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Build of instance c5aaadb0-f9a1-4af9-b882-86d847185a19 was re-scheduled: Binding failed for port 83e791da-9e3a-4c0f-9454-6662da4e8798, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 697.757901] env[62914]: DEBUG nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 697.758145] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "refresh_cache-c5aaadb0-f9a1-4af9-b882-86d847185a19" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.758292] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "refresh_cache-c5aaadb0-f9a1-4af9-b882-86d847185a19" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.758453] env[62914]: DEBUG nova.network.neutron [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.759841] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.895s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.761607] env[62914]: INFO nova.compute.claims [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.767433] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Releasing lock "refresh_cache-ad283bc7-b422-4836-a503-d3d9ae4fe3c1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.767788] env[62914]: DEBUG nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 697.767972] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 697.768476] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f264eefb-a1eb-4abb-882c-2b3f5f7bc334 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.778843] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c2da7eb-7fdb-4426-ab3d-c2c87122b81d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.800562] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ad283bc7-b422-4836-a503-d3d9ae4fe3c1 could not be found. [ 697.800768] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 697.800942] env[62914]: INFO nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 697.801192] env[62914]: DEBUG oslo.service.loopingcall [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 697.801386] env[62914]: DEBUG nova.compute.manager [-] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 697.801481] env[62914]: DEBUG nova.network.neutron [-] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 697.826019] env[62914]: DEBUG nova.network.neutron [-] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.007344] env[62914]: INFO nova.compute.manager [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] [instance: c10d840e-ce05-4af3-ba2d-6214eefd8783] Took 1.02 seconds to deallocate network for instance. [ 698.285556] env[62914]: DEBUG nova.network.neutron [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.327650] env[62914]: DEBUG nova.network.neutron [-] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.373887] env[62914]: DEBUG nova.network.neutron [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.831499] env[62914]: INFO nova.compute.manager [-] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Took 1.03 seconds to deallocate network for instance. [ 698.833828] env[62914]: DEBUG nova.compute.claims [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 698.834656] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.876614] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "refresh_cache-c5aaadb0-f9a1-4af9-b882-86d847185a19" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.876840] env[62914]: DEBUG nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 698.877025] env[62914]: DEBUG nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.877186] env[62914]: DEBUG nova.network.neutron [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 698.891557] env[62914]: DEBUG nova.network.neutron [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.036761] env[62914]: INFO nova.scheduler.client.report [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Deleted allocations for instance c10d840e-ce05-4af3-ba2d-6214eefd8783 [ 699.089697] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9abfe8d4-f236-4a17-b6e5-d34f973579b8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.098015] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ca965e-fd23-44df-9f0e-de9d4d3a23a1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.128019] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c455d3e3-c019-4fbf-b505-2c980ee41c3d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.135293] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd49f0d-f178-416f-8860-1ec6a1d68845 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.149483] env[62914]: DEBUG nova.compute.provider_tree [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 699.394701] env[62914]: DEBUG nova.network.neutron [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.548547] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7177ce71-cb29-448a-bd7d-e72d0fd895fa tempest-ServersTestFqdnHostnames-258941681 tempest-ServersTestFqdnHostnames-258941681-project-member] Lock "c10d840e-ce05-4af3-ba2d-6214eefd8783" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.076s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.654315] env[62914]: DEBUG nova.scheduler.client.report [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.897251] env[62914]: INFO nova.compute.manager [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: c5aaadb0-f9a1-4af9-b882-86d847185a19] Took 1.02 seconds to deallocate network for instance. [ 700.051065] env[62914]: DEBUG nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 700.159165] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.159665] env[62914]: DEBUG nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 700.162074] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.627s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.582376] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.671438] env[62914]: DEBUG nova.compute.utils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 700.672251] env[62914]: DEBUG nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 700.672686] env[62914]: DEBUG nova.network.neutron [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 700.737477] env[62914]: DEBUG nova.policy [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50bf8e4690b841d6ac4f49888b0b3e64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8fdabf0b14c244f5b7bba48f9ef5d941', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 700.938571] env[62914]: INFO nova.scheduler.client.report [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Deleted allocations for instance c5aaadb0-f9a1-4af9-b882-86d847185a19 [ 701.034751] env[62914]: DEBUG nova.network.neutron [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Successfully created port: 7bcf9042-0227-4463-ae3a-769768b437a5 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 701.044962] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e077c355-a14b-41d4-a3b0-230a214e1583 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.053970] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f88c56d-317d-4c73-bff4-59aa6d3affc6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.086407] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd55fc4-12ed-4b7a-8e16-b0f96c6fb3d4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.093856] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed30866-3dc4-49f9-868f-c9d03a289925 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.109118] env[62914]: DEBUG nova.compute.provider_tree [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.177017] env[62914]: DEBUG nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 701.450318] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d929aaf4-cfd3-45c8-a633-8aaba699ad2b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "c5aaadb0-f9a1-4af9-b882-86d847185a19" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.492s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.611904] env[62914]: DEBUG nova.scheduler.client.report [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.957739] env[62914]: DEBUG nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 702.120385] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.958s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.120999] env[62914]: ERROR nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8471114f-49d5-4173-90f1-d847bc66942f, please check neutron logs for more information. [ 702.120999] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Traceback (most recent call last): [ 702.120999] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.120999] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] self.driver.spawn(context, instance, image_meta, [ 702.120999] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 702.120999] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.120999] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.120999] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] vm_ref = self.build_virtual_machine(instance, [ 702.120999] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.120999] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.120999] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.121325] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] for vif in network_info: [ 702.121325] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 702.121325] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] return self._sync_wrapper(fn, *args, **kwargs) [ 702.121325] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 702.121325] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] self.wait() [ 702.121325] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 702.121325] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] self[:] = self._gt.wait() [ 702.121325] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.121325] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] return self._exit_event.wait() [ 702.121325] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 702.121325] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] current.throw(*self._exc) [ 702.121325] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.121325] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] result = function(*args, **kwargs) [ 702.121733] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 702.121733] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] return func(*args, **kwargs) [ 702.121733] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.121733] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] raise e [ 702.121733] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.121733] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] nwinfo = self.network_api.allocate_for_instance( [ 702.121733] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.121733] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] created_port_ids = self._update_ports_for_instance( [ 702.121733] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.121733] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] with excutils.save_and_reraise_exception(): [ 702.121733] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.121733] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] self.force_reraise() [ 702.121733] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.122071] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] raise self.value [ 702.122071] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.122071] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] updated_port = self._update_port( [ 702.122071] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.122071] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] _ensure_no_port_binding_failure(port) [ 702.122071] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.122071] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] raise exception.PortBindingFailed(port_id=port['id']) [ 702.122071] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] nova.exception.PortBindingFailed: Binding failed for port 8471114f-49d5-4173-90f1-d847bc66942f, please check neutron logs for more information. [ 702.122071] env[62914]: ERROR nova.compute.manager [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] [ 702.122071] env[62914]: DEBUG nova.compute.utils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Binding failed for port 8471114f-49d5-4173-90f1-d847bc66942f, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 702.124912] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.923s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.127250] env[62914]: INFO nova.compute.claims [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 702.130947] env[62914]: DEBUG nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Build of instance 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089 was re-scheduled: Binding failed for port 8471114f-49d5-4173-90f1-d847bc66942f, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 702.131613] env[62914]: DEBUG nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 702.131736] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Acquiring lock "refresh_cache-62cfca0a-e8c6-4f94-8e6f-1136cfcf3089" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.131824] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Acquired lock "refresh_cache-62cfca0a-e8c6-4f94-8e6f-1136cfcf3089" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.132039] env[62914]: DEBUG nova.network.neutron [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.188734] env[62914]: DEBUG nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 702.214206] env[62914]: DEBUG nova.compute.manager [req-8144fdb7-916a-44cb-968b-18e520d0ad5e req-d9ba2cdc-0874-44dd-a126-5b47e80aa5b3 service nova] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Received event network-changed-7bcf9042-0227-4463-ae3a-769768b437a5 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 702.214462] env[62914]: DEBUG nova.compute.manager [req-8144fdb7-916a-44cb-968b-18e520d0ad5e req-d9ba2cdc-0874-44dd-a126-5b47e80aa5b3 service nova] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Refreshing instance network info cache due to event network-changed-7bcf9042-0227-4463-ae3a-769768b437a5. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 702.214706] env[62914]: DEBUG oslo_concurrency.lockutils [req-8144fdb7-916a-44cb-968b-18e520d0ad5e req-d9ba2cdc-0874-44dd-a126-5b47e80aa5b3 service nova] Acquiring lock "refresh_cache-67554dc7-fb72-47e1-8201-f5fda035a2d3" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.218193] env[62914]: DEBUG oslo_concurrency.lockutils [req-8144fdb7-916a-44cb-968b-18e520d0ad5e req-d9ba2cdc-0874-44dd-a126-5b47e80aa5b3 service nova] Acquired lock "refresh_cache-67554dc7-fb72-47e1-8201-f5fda035a2d3" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.218193] env[62914]: DEBUG nova.network.neutron [req-8144fdb7-916a-44cb-968b-18e520d0ad5e req-d9ba2cdc-0874-44dd-a126-5b47e80aa5b3 service nova] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Refreshing network info cache for port 7bcf9042-0227-4463-ae3a-769768b437a5 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 702.232679] env[62914]: DEBUG nova.virt.hardware [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 702.232991] env[62914]: DEBUG nova.virt.hardware [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 702.233250] env[62914]: DEBUG nova.virt.hardware [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 702.233488] env[62914]: DEBUG nova.virt.hardware [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 702.233687] env[62914]: DEBUG nova.virt.hardware [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 702.233917] env[62914]: DEBUG nova.virt.hardware [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 702.234188] env[62914]: DEBUG nova.virt.hardware [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 702.234427] env[62914]: DEBUG nova.virt.hardware [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 702.234663] env[62914]: DEBUG nova.virt.hardware [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 702.234882] env[62914]: DEBUG nova.virt.hardware [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 702.235396] env[62914]: DEBUG nova.virt.hardware [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 702.236151] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb11b3d-4f3c-435f-bc4a-c27a3ce76613 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.247309] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe125478-f74c-46fb-859e-66d5f0863336 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.353797] env[62914]: ERROR nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7bcf9042-0227-4463-ae3a-769768b437a5, please check neutron logs for more information. [ 702.353797] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 702.353797] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.353797] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 702.353797] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.353797] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 702.353797] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.353797] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 702.353797] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.353797] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 702.353797] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.353797] env[62914]: ERROR nova.compute.manager raise self.value [ 702.353797] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.353797] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 702.353797] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.353797] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 702.354641] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.354641] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 702.354641] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7bcf9042-0227-4463-ae3a-769768b437a5, please check neutron logs for more information. [ 702.354641] env[62914]: ERROR nova.compute.manager [ 702.354641] env[62914]: Traceback (most recent call last): [ 702.354641] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 702.354641] env[62914]: listener.cb(fileno) [ 702.354641] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.354641] env[62914]: result = function(*args, **kwargs) [ 702.354641] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 702.354641] env[62914]: return func(*args, **kwargs) [ 702.354641] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.354641] env[62914]: raise e [ 702.354641] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.354641] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 702.354641] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.354641] env[62914]: created_port_ids = self._update_ports_for_instance( [ 702.354641] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.354641] env[62914]: with excutils.save_and_reraise_exception(): [ 702.354641] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.354641] env[62914]: self.force_reraise() [ 702.354641] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.354641] env[62914]: raise self.value [ 702.354641] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.354641] env[62914]: updated_port = self._update_port( [ 702.354641] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.354641] env[62914]: _ensure_no_port_binding_failure(port) [ 702.354641] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.354641] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 702.355884] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 7bcf9042-0227-4463-ae3a-769768b437a5, please check neutron logs for more information. [ 702.355884] env[62914]: Removing descriptor: 15 [ 702.355884] env[62914]: ERROR nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7bcf9042-0227-4463-ae3a-769768b437a5, please check neutron logs for more information. [ 702.355884] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Traceback (most recent call last): [ 702.355884] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 702.355884] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] yield resources [ 702.355884] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.355884] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] self.driver.spawn(context, instance, image_meta, [ 702.355884] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 702.355884] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.355884] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.355884] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] vm_ref = self.build_virtual_machine(instance, [ 702.356500] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.356500] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.356500] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.356500] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] for vif in network_info: [ 702.356500] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 702.356500] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] return self._sync_wrapper(fn, *args, **kwargs) [ 702.356500] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 702.356500] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] self.wait() [ 702.356500] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 702.356500] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] self[:] = self._gt.wait() [ 702.356500] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.356500] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] return self._exit_event.wait() [ 702.356500] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 702.356929] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] result = hub.switch() [ 702.356929] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 702.356929] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] return self.greenlet.switch() [ 702.356929] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.356929] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] result = function(*args, **kwargs) [ 702.356929] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 702.356929] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] return func(*args, **kwargs) [ 702.356929] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.356929] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] raise e [ 702.356929] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.356929] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] nwinfo = self.network_api.allocate_for_instance( [ 702.356929] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.356929] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] created_port_ids = self._update_ports_for_instance( [ 702.357281] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.357281] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] with excutils.save_and_reraise_exception(): [ 702.357281] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.357281] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] self.force_reraise() [ 702.357281] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.357281] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] raise self.value [ 702.357281] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.357281] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] updated_port = self._update_port( [ 702.357281] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.357281] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] _ensure_no_port_binding_failure(port) [ 702.357281] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.357281] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] raise exception.PortBindingFailed(port_id=port['id']) [ 702.357779] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] nova.exception.PortBindingFailed: Binding failed for port 7bcf9042-0227-4463-ae3a-769768b437a5, please check neutron logs for more information. [ 702.357779] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] [ 702.357779] env[62914]: INFO nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Terminating instance [ 702.357779] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Acquiring lock "refresh_cache-67554dc7-fb72-47e1-8201-f5fda035a2d3" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.478106] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.575456] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "6e38d30c-2fa3-468d-8ff9-e3966333c9c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.575706] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "6e38d30c-2fa3-468d-8ff9-e3966333c9c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.656238] env[62914]: DEBUG nova.network.neutron [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.742804] env[62914]: DEBUG nova.network.neutron [req-8144fdb7-916a-44cb-968b-18e520d0ad5e req-d9ba2cdc-0874-44dd-a126-5b47e80aa5b3 service nova] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.761755] env[62914]: DEBUG nova.network.neutron [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.832584] env[62914]: DEBUG nova.network.neutron [req-8144fdb7-916a-44cb-968b-18e520d0ad5e req-d9ba2cdc-0874-44dd-a126-5b47e80aa5b3 service nova] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.265679] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Releasing lock "refresh_cache-62cfca0a-e8c6-4f94-8e6f-1136cfcf3089" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.265964] env[62914]: DEBUG nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 703.266685] env[62914]: DEBUG nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.266685] env[62914]: DEBUG nova.network.neutron [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 703.289039] env[62914]: DEBUG nova.network.neutron [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.335766] env[62914]: DEBUG oslo_concurrency.lockutils [req-8144fdb7-916a-44cb-968b-18e520d0ad5e req-d9ba2cdc-0874-44dd-a126-5b47e80aa5b3 service nova] Releasing lock "refresh_cache-67554dc7-fb72-47e1-8201-f5fda035a2d3" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.336203] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Acquired lock "refresh_cache-67554dc7-fb72-47e1-8201-f5fda035a2d3" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.336390] env[62914]: DEBUG nova.network.neutron [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 703.494322] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e95de6dd-604c-4e68-95de-d9e9c5013514 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.501998] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf463c7b-d977-4737-b612-7aee114461db {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.533080] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b641c55b-9d06-4fa0-bd3d-8091453f4012 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.541483] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044efbfa-89c5-4059-a713-8f959569e4c7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.555712] env[62914]: DEBUG nova.compute.provider_tree [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.792770] env[62914]: DEBUG nova.network.neutron [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.853481] env[62914]: DEBUG nova.network.neutron [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.935073] env[62914]: DEBUG nova.network.neutron [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.059564] env[62914]: DEBUG nova.scheduler.client.report [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 704.236585] env[62914]: DEBUG nova.compute.manager [req-c76236a1-c565-45f1-aafc-40fbed336179 req-be3e0933-193b-4f68-a894-99876bad36e7 service nova] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Received event network-vif-deleted-7bcf9042-0227-4463-ae3a-769768b437a5 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.294939] env[62914]: INFO nova.compute.manager [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] [instance: 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089] Took 1.03 seconds to deallocate network for instance. [ 704.437544] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Releasing lock "refresh_cache-67554dc7-fb72-47e1-8201-f5fda035a2d3" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.438173] env[62914]: DEBUG nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 704.438460] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 704.438782] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-86298a93-6419-4408-9c08-fd32431f4993 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.448186] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e814620-33f2-4aa3-8e78-4103c19aa2e1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.468368] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 67554dc7-fb72-47e1-8201-f5fda035a2d3 could not be found. [ 704.468551] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 704.468724] env[62914]: INFO nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 704.468977] env[62914]: DEBUG oslo.service.loopingcall [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 704.469168] env[62914]: DEBUG nova.compute.manager [-] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.469262] env[62914]: DEBUG nova.network.neutron [-] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 704.483575] env[62914]: DEBUG nova.network.neutron [-] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.565115] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.565115] env[62914]: DEBUG nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 704.568211] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.272s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.985383] env[62914]: DEBUG nova.network.neutron [-] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.073059] env[62914]: DEBUG nova.compute.utils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 705.077057] env[62914]: DEBUG nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 705.077057] env[62914]: DEBUG nova.network.neutron [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 705.116179] env[62914]: DEBUG nova.policy [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '19afcb6d7c204ccba0018a1481baf0cf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cb44d9a5537a4a9497c83c0ebe68c521', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 705.326673] env[62914]: INFO nova.scheduler.client.report [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Deleted allocations for instance 62cfca0a-e8c6-4f94-8e6f-1136cfcf3089 [ 705.431909] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae48f20-7e1f-41e9-b62e-6970ef39c8f0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.439617] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b4ac9a-04c2-4fc6-8986-23b7c4e4c48b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.475351] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25c967c-3c4b-444b-b712-ea597b61203b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.483710] env[62914]: DEBUG nova.network.neutron [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Successfully created port: e5560f64-e00b-4658-87c5-b88d50d90135 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.486436] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d537092-d3eb-4281-8f5c-9a46be817d44 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.490978] env[62914]: INFO nova.compute.manager [-] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Took 1.02 seconds to deallocate network for instance. [ 705.493523] env[62914]: DEBUG nova.compute.claims [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 705.493644] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.501927] env[62914]: DEBUG nova.compute.provider_tree [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.577502] env[62914]: DEBUG nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 705.844020] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7c0507f8-46dc-4e75-9cc3-dc7ae6dba4bf tempest-ServersTestJSON-403245787 tempest-ServersTestJSON-403245787-project-member] Lock "62cfca0a-e8c6-4f94-8e6f-1136cfcf3089" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.413s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.006396] env[62914]: DEBUG nova.scheduler.client.report [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.288297] env[62914]: DEBUG nova.compute.manager [req-b1a9c22c-644c-4cfd-8cc3-9fb1abffe58e req-5756acdb-3719-4129-8cf1-d58d2b9b9789 service nova] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Received event network-changed-e5560f64-e00b-4658-87c5-b88d50d90135 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 706.288494] env[62914]: DEBUG nova.compute.manager [req-b1a9c22c-644c-4cfd-8cc3-9fb1abffe58e req-5756acdb-3719-4129-8cf1-d58d2b9b9789 service nova] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Refreshing instance network info cache due to event network-changed-e5560f64-e00b-4658-87c5-b88d50d90135. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 706.288697] env[62914]: DEBUG oslo_concurrency.lockutils [req-b1a9c22c-644c-4cfd-8cc3-9fb1abffe58e req-5756acdb-3719-4129-8cf1-d58d2b9b9789 service nova] Acquiring lock "refresh_cache-821366f1-f1b0-4b78-8f7e-db8fd3bb123e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.288842] env[62914]: DEBUG oslo_concurrency.lockutils [req-b1a9c22c-644c-4cfd-8cc3-9fb1abffe58e req-5756acdb-3719-4129-8cf1-d58d2b9b9789 service nova] Acquired lock "refresh_cache-821366f1-f1b0-4b78-8f7e-db8fd3bb123e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.288995] env[62914]: DEBUG nova.network.neutron [req-b1a9c22c-644c-4cfd-8cc3-9fb1abffe58e req-5756acdb-3719-4129-8cf1-d58d2b9b9789 service nova] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Refreshing network info cache for port e5560f64-e00b-4658-87c5-b88d50d90135 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 706.350018] env[62914]: DEBUG nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 706.502952] env[62914]: ERROR nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e5560f64-e00b-4658-87c5-b88d50d90135, please check neutron logs for more information. [ 706.502952] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 706.502952] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.502952] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 706.502952] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.502952] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 706.502952] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.502952] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 706.502952] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.502952] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 706.502952] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.502952] env[62914]: ERROR nova.compute.manager raise self.value [ 706.502952] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.502952] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 706.502952] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.502952] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 706.503639] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.503639] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 706.503639] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e5560f64-e00b-4658-87c5-b88d50d90135, please check neutron logs for more information. [ 706.503639] env[62914]: ERROR nova.compute.manager [ 706.503639] env[62914]: Traceback (most recent call last): [ 706.503639] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 706.503639] env[62914]: listener.cb(fileno) [ 706.503639] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.503639] env[62914]: result = function(*args, **kwargs) [ 706.503639] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 706.503639] env[62914]: return func(*args, **kwargs) [ 706.503639] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.503639] env[62914]: raise e [ 706.503639] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.503639] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 706.503639] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.503639] env[62914]: created_port_ids = self._update_ports_for_instance( [ 706.503639] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.503639] env[62914]: with excutils.save_and_reraise_exception(): [ 706.503639] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.503639] env[62914]: self.force_reraise() [ 706.503639] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.503639] env[62914]: raise self.value [ 706.503639] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.503639] env[62914]: updated_port = self._update_port( [ 706.503639] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.503639] env[62914]: _ensure_no_port_binding_failure(port) [ 706.503639] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.503639] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 706.508562] env[62914]: nova.exception.PortBindingFailed: Binding failed for port e5560f64-e00b-4658-87c5-b88d50d90135, please check neutron logs for more information. [ 706.508562] env[62914]: Removing descriptor: 19 [ 706.510847] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.943s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.511462] env[62914]: ERROR nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8881157e-8fc3-480e-be24-9ea26a9df050, please check neutron logs for more information. [ 706.511462] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] Traceback (most recent call last): [ 706.511462] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.511462] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] self.driver.spawn(context, instance, image_meta, [ 706.511462] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 706.511462] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.511462] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.511462] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] vm_ref = self.build_virtual_machine(instance, [ 706.511462] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.511462] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.511462] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.511951] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] for vif in network_info: [ 706.511951] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.511951] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] return self._sync_wrapper(fn, *args, **kwargs) [ 706.511951] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.511951] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] self.wait() [ 706.511951] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.511951] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] self[:] = self._gt.wait() [ 706.511951] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.511951] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] return self._exit_event.wait() [ 706.511951] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 706.511951] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] result = hub.switch() [ 706.511951] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 706.511951] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] return self.greenlet.switch() [ 706.512709] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.512709] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] result = function(*args, **kwargs) [ 706.512709] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 706.512709] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] return func(*args, **kwargs) [ 706.512709] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.512709] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] raise e [ 706.512709] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.512709] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] nwinfo = self.network_api.allocate_for_instance( [ 706.512709] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.512709] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] created_port_ids = self._update_ports_for_instance( [ 706.512709] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.512709] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] with excutils.save_and_reraise_exception(): [ 706.512709] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.513200] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] self.force_reraise() [ 706.513200] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.513200] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] raise self.value [ 706.513200] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.513200] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] updated_port = self._update_port( [ 706.513200] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.513200] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] _ensure_no_port_binding_failure(port) [ 706.513200] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.513200] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] raise exception.PortBindingFailed(port_id=port['id']) [ 706.513200] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] nova.exception.PortBindingFailed: Binding failed for port 8881157e-8fc3-480e-be24-9ea26a9df050, please check neutron logs for more information. [ 706.513200] env[62914]: ERROR nova.compute.manager [instance: 294e3321-e826-4a56-bbd0-893261d65283] [ 706.513970] env[62914]: DEBUG nova.compute.utils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Binding failed for port 8881157e-8fc3-480e-be24-9ea26a9df050, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 706.514095] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.392s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.519833] env[62914]: DEBUG nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Build of instance 294e3321-e826-4a56-bbd0-893261d65283 was re-scheduled: Binding failed for port 8881157e-8fc3-480e-be24-9ea26a9df050, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 706.519833] env[62914]: DEBUG nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 706.519833] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Acquiring lock "refresh_cache-294e3321-e826-4a56-bbd0-893261d65283" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.520077] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Acquired lock "refresh_cache-294e3321-e826-4a56-bbd0-893261d65283" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.520134] env[62914]: DEBUG nova.network.neutron [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 706.591467] env[62914]: DEBUG nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 706.623029] env[62914]: DEBUG nova.virt.hardware [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 706.627613] env[62914]: DEBUG nova.virt.hardware [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 706.627613] env[62914]: DEBUG nova.virt.hardware [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.627613] env[62914]: DEBUG nova.virt.hardware [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 706.627814] env[62914]: DEBUG nova.virt.hardware [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.627963] env[62914]: DEBUG nova.virt.hardware [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 706.628204] env[62914]: DEBUG nova.virt.hardware [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 706.628356] env[62914]: DEBUG nova.virt.hardware [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 706.628515] env[62914]: DEBUG nova.virt.hardware [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 706.628711] env[62914]: DEBUG nova.virt.hardware [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 706.628913] env[62914]: DEBUG nova.virt.hardware [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.630203] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de34a64-554d-4e79-ad31-cc997bd17651 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.639560] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb4261c6-1021-403d-aa6f-f8c7615c07b1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.653730] env[62914]: ERROR nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e5560f64-e00b-4658-87c5-b88d50d90135, please check neutron logs for more information. [ 706.653730] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Traceback (most recent call last): [ 706.653730] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 706.653730] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] yield resources [ 706.653730] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.653730] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] self.driver.spawn(context, instance, image_meta, [ 706.653730] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 706.653730] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.653730] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.653730] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] vm_ref = self.build_virtual_machine(instance, [ 706.653730] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.654135] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.654135] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.654135] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] for vif in network_info: [ 706.654135] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.654135] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] return self._sync_wrapper(fn, *args, **kwargs) [ 706.654135] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.654135] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] self.wait() [ 706.654135] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.654135] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] self[:] = self._gt.wait() [ 706.654135] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.654135] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] return self._exit_event.wait() [ 706.654135] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 706.654135] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] current.throw(*self._exc) [ 706.654510] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.654510] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] result = function(*args, **kwargs) [ 706.654510] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 706.654510] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] return func(*args, **kwargs) [ 706.654510] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.654510] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] raise e [ 706.654510] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.654510] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] nwinfo = self.network_api.allocate_for_instance( [ 706.654510] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.654510] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] created_port_ids = self._update_ports_for_instance( [ 706.654510] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.654510] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] with excutils.save_and_reraise_exception(): [ 706.654510] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.654862] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] self.force_reraise() [ 706.654862] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.654862] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] raise self.value [ 706.654862] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.654862] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] updated_port = self._update_port( [ 706.654862] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.654862] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] _ensure_no_port_binding_failure(port) [ 706.654862] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.654862] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] raise exception.PortBindingFailed(port_id=port['id']) [ 706.654862] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] nova.exception.PortBindingFailed: Binding failed for port e5560f64-e00b-4658-87c5-b88d50d90135, please check neutron logs for more information. [ 706.654862] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] [ 706.654862] env[62914]: INFO nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Terminating instance [ 706.656593] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Acquiring lock "refresh_cache-821366f1-f1b0-4b78-8f7e-db8fd3bb123e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.808078] env[62914]: DEBUG nova.network.neutron [req-b1a9c22c-644c-4cfd-8cc3-9fb1abffe58e req-5756acdb-3719-4129-8cf1-d58d2b9b9789 service nova] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.876794] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.899081] env[62914]: DEBUG nova.network.neutron [req-b1a9c22c-644c-4cfd-8cc3-9fb1abffe58e req-5756acdb-3719-4129-8cf1-d58d2b9b9789 service nova] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.044194] env[62914]: DEBUG nova.network.neutron [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.144746] env[62914]: DEBUG nova.network.neutron [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.360068] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09418d0-ef48-4ca8-baff-fa792a1491b3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.367873] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf42aa6-181e-48d9-9f49-8e28d975eefd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.398360] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d8c9a9-f866-422d-a941-38557448e4a6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.401211] env[62914]: DEBUG oslo_concurrency.lockutils [req-b1a9c22c-644c-4cfd-8cc3-9fb1abffe58e req-5756acdb-3719-4129-8cf1-d58d2b9b9789 service nova] Releasing lock "refresh_cache-821366f1-f1b0-4b78-8f7e-db8fd3bb123e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.401569] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Acquired lock "refresh_cache-821366f1-f1b0-4b78-8f7e-db8fd3bb123e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.401753] env[62914]: DEBUG nova.network.neutron [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.407811] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91b066d-88f9-4a7c-a784-ff52b5d99c45 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.421941] env[62914]: DEBUG nova.compute.provider_tree [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.649986] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Releasing lock "refresh_cache-294e3321-e826-4a56-bbd0-893261d65283" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.650132] env[62914]: DEBUG nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 707.650366] env[62914]: DEBUG nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.650608] env[62914]: DEBUG nova.network.neutron [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 707.668441] env[62914]: DEBUG nova.network.neutron [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.921783] env[62914]: DEBUG nova.network.neutron [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.924077] env[62914]: DEBUG nova.scheduler.client.report [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.049627] env[62914]: DEBUG nova.network.neutron [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.170655] env[62914]: DEBUG nova.network.neutron [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.314872] env[62914]: DEBUG nova.compute.manager [req-65b8f48d-f2e6-4783-9399-0ae52a6211b2 req-58a3501d-ee91-4efe-b1b9-4fe14679e5bd service nova] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Received event network-vif-deleted-e5560f64-e00b-4658-87c5-b88d50d90135 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.429420] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.916s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.430122] env[62914]: ERROR nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 25685cb7-8693-45a1-8d41-212a292f6db9, please check neutron logs for more information. [ 708.430122] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Traceback (most recent call last): [ 708.430122] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 708.430122] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] self.driver.spawn(context, instance, image_meta, [ 708.430122] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 708.430122] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 708.430122] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 708.430122] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] vm_ref = self.build_virtual_machine(instance, [ 708.430122] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 708.430122] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] vif_infos = vmwarevif.get_vif_info(self._session, [ 708.430122] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 708.430512] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] for vif in network_info: [ 708.430512] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 708.430512] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] return self._sync_wrapper(fn, *args, **kwargs) [ 708.430512] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 708.430512] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] self.wait() [ 708.430512] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 708.430512] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] self[:] = self._gt.wait() [ 708.430512] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 708.430512] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] return self._exit_event.wait() [ 708.430512] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 708.430512] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] result = hub.switch() [ 708.430512] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 708.430512] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] return self.greenlet.switch() [ 708.430842] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.430842] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] result = function(*args, **kwargs) [ 708.430842] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 708.430842] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] return func(*args, **kwargs) [ 708.430842] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.430842] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] raise e [ 708.430842] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.430842] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] nwinfo = self.network_api.allocate_for_instance( [ 708.430842] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.430842] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] created_port_ids = self._update_ports_for_instance( [ 708.430842] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.430842] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] with excutils.save_and_reraise_exception(): [ 708.430842] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.431180] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] self.force_reraise() [ 708.431180] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.431180] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] raise self.value [ 708.431180] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.431180] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] updated_port = self._update_port( [ 708.431180] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.431180] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] _ensure_no_port_binding_failure(port) [ 708.431180] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.431180] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] raise exception.PortBindingFailed(port_id=port['id']) [ 708.431180] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] nova.exception.PortBindingFailed: Binding failed for port 25685cb7-8693-45a1-8d41-212a292f6db9, please check neutron logs for more information. [ 708.431180] env[62914]: ERROR nova.compute.manager [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] [ 708.431461] env[62914]: DEBUG nova.compute.utils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Binding failed for port 25685cb7-8693-45a1-8d41-212a292f6db9, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 708.432160] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.629s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.435338] env[62914]: DEBUG nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Build of instance 0ac803ee-6303-463f-91e8-8263c0219b0a was re-scheduled: Binding failed for port 25685cb7-8693-45a1-8d41-212a292f6db9, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 708.435423] env[62914]: DEBUG nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 708.436076] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Acquiring lock "refresh_cache-0ac803ee-6303-463f-91e8-8263c0219b0a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.436076] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Acquired lock "refresh_cache-0ac803ee-6303-463f-91e8-8263c0219b0a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.436076] env[62914]: DEBUG nova.network.neutron [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 708.553022] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Releasing lock "refresh_cache-821366f1-f1b0-4b78-8f7e-db8fd3bb123e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.553022] env[62914]: DEBUG nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 708.553022] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 708.553352] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d06ca20-6c51-4404-a9e6-6f66078fc425 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.563167] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699692c8-4b77-4240-abb8-d28ebb9888bf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.584331] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 821366f1-f1b0-4b78-8f7e-db8fd3bb123e could not be found. [ 708.584532] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 708.584735] env[62914]: INFO nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 708.585779] env[62914]: DEBUG oslo.service.loopingcall [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.585779] env[62914]: DEBUG nova.compute.manager [-] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.585779] env[62914]: DEBUG nova.network.neutron [-] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.600861] env[62914]: DEBUG nova.network.neutron [-] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.675997] env[62914]: INFO nova.compute.manager [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 294e3321-e826-4a56-bbd0-893261d65283] Took 1.03 seconds to deallocate network for instance. [ 708.955945] env[62914]: DEBUG nova.network.neutron [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.103869] env[62914]: DEBUG nova.network.neutron [-] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.171853] env[62914]: DEBUG nova.network.neutron [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.288028] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff8bf58-dd5e-4133-a4de-76fcc2e71822 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.294074] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dcec3b0-91ad-4b2b-9516-e8fb763713b9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.324691] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f7174b-c441-4fbb-9571-6b2895368a22 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.331429] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9db790a-4e04-4dcc-a890-e2e2e87265fc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.344789] env[62914]: DEBUG nova.compute.provider_tree [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.606545] env[62914]: INFO nova.compute.manager [-] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Took 1.02 seconds to deallocate network for instance. [ 709.609117] env[62914]: DEBUG nova.compute.claims [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 709.609323] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.674965] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Releasing lock "refresh_cache-0ac803ee-6303-463f-91e8-8263c0219b0a" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.674965] env[62914]: DEBUG nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 709.675209] env[62914]: DEBUG nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 709.675245] env[62914]: DEBUG nova.network.neutron [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 709.690171] env[62914]: DEBUG nova.network.neutron [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.708027] env[62914]: INFO nova.scheduler.client.report [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Deleted allocations for instance 294e3321-e826-4a56-bbd0-893261d65283 [ 709.847882] env[62914]: DEBUG nova.scheduler.client.report [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.192992] env[62914]: DEBUG nova.network.neutron [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.216146] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5d9e9405-5ce5-403c-ae00-1901fb9cf06e tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Lock "294e3321-e826-4a56-bbd0-893261d65283" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.999s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.352597] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.920s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.353096] env[62914]: ERROR nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 26984a2f-dd05-4896-9dc6-c35d18f01310, please check neutron logs for more information. [ 710.353096] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Traceback (most recent call last): [ 710.353096] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.353096] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] self.driver.spawn(context, instance, image_meta, [ 710.353096] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 710.353096] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.353096] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.353096] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] vm_ref = self.build_virtual_machine(instance, [ 710.353096] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.353096] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.353096] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.353421] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] for vif in network_info: [ 710.353421] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.353421] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] return self._sync_wrapper(fn, *args, **kwargs) [ 710.353421] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.353421] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] self.wait() [ 710.353421] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.353421] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] self[:] = self._gt.wait() [ 710.353421] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.353421] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] return self._exit_event.wait() [ 710.353421] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 710.353421] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] result = hub.switch() [ 710.353421] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 710.353421] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] return self.greenlet.switch() [ 710.353823] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.353823] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] result = function(*args, **kwargs) [ 710.353823] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 710.353823] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] return func(*args, **kwargs) [ 710.353823] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.353823] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] raise e [ 710.353823] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.353823] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] nwinfo = self.network_api.allocate_for_instance( [ 710.353823] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.353823] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] created_port_ids = self._update_ports_for_instance( [ 710.353823] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.353823] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] with excutils.save_and_reraise_exception(): [ 710.353823] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.354192] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] self.force_reraise() [ 710.354192] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.354192] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] raise self.value [ 710.354192] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.354192] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] updated_port = self._update_port( [ 710.354192] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.354192] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] _ensure_no_port_binding_failure(port) [ 710.354192] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.354192] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] raise exception.PortBindingFailed(port_id=port['id']) [ 710.354192] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] nova.exception.PortBindingFailed: Binding failed for port 26984a2f-dd05-4896-9dc6-c35d18f01310, please check neutron logs for more information. [ 710.354192] env[62914]: ERROR nova.compute.manager [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] [ 710.354508] env[62914]: DEBUG nova.compute.utils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Binding failed for port 26984a2f-dd05-4896-9dc6-c35d18f01310, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 710.355238] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.728s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.359918] env[62914]: DEBUG nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Build of instance 3ce62ec7-6274-441f-989b-6f27a64b0c18 was re-scheduled: Binding failed for port 26984a2f-dd05-4896-9dc6-c35d18f01310, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 710.360393] env[62914]: DEBUG nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 710.360614] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Acquiring lock "refresh_cache-3ce62ec7-6274-441f-989b-6f27a64b0c18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.360750] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Acquired lock "refresh_cache-3ce62ec7-6274-441f-989b-6f27a64b0c18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.361331] env[62914]: DEBUG nova.network.neutron [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 710.696947] env[62914]: INFO nova.compute.manager [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 0ac803ee-6303-463f-91e8-8263c0219b0a] Took 1.02 seconds to deallocate network for instance. [ 710.718424] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 710.885103] env[62914]: DEBUG nova.network.neutron [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.937128] env[62914]: DEBUG nova.network.neutron [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.185737] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a06aac1b-9867-454e-b0a3-2a9bdf2a5baf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.193152] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c8ab39-4a37-4bbd-b64f-48f7c5e85996 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.232789] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb9fb5b-7210-46cf-9d42-043fb43463d4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.240410] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa84b7f-8b6f-4ea5-83a1-ddf4057dcc5b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.254110] env[62914]: DEBUG nova.compute.provider_tree [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.256679] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.444292] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Releasing lock "refresh_cache-3ce62ec7-6274-441f-989b-6f27a64b0c18" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.444292] env[62914]: DEBUG nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 711.444292] env[62914]: DEBUG nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.444292] env[62914]: DEBUG nova.network.neutron [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 711.461044] env[62914]: DEBUG nova.network.neutron [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.758282] env[62914]: DEBUG nova.scheduler.client.report [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.768038] env[62914]: INFO nova.scheduler.client.report [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Deleted allocations for instance 0ac803ee-6303-463f-91e8-8263c0219b0a [ 711.964222] env[62914]: DEBUG nova.network.neutron [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.262982] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.908s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.263748] env[62914]: ERROR nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a6637332-9313-492b-bdde-1386cdbb1191, please check neutron logs for more information. [ 712.263748] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Traceback (most recent call last): [ 712.263748] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.263748] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] self.driver.spawn(context, instance, image_meta, [ 712.263748] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 712.263748] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.263748] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.263748] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] vm_ref = self.build_virtual_machine(instance, [ 712.263748] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.263748] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.263748] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.264198] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] for vif in network_info: [ 712.264198] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.264198] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] return self._sync_wrapper(fn, *args, **kwargs) [ 712.264198] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.264198] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] self.wait() [ 712.264198] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.264198] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] self[:] = self._gt.wait() [ 712.264198] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.264198] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] return self._exit_event.wait() [ 712.264198] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 712.264198] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] current.throw(*self._exc) [ 712.264198] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.264198] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] result = function(*args, **kwargs) [ 712.264626] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.264626] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] return func(*args, **kwargs) [ 712.264626] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.264626] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] raise e [ 712.264626] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.264626] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] nwinfo = self.network_api.allocate_for_instance( [ 712.264626] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.264626] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] created_port_ids = self._update_ports_for_instance( [ 712.264626] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.264626] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] with excutils.save_and_reraise_exception(): [ 712.264626] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.264626] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] self.force_reraise() [ 712.264626] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.265109] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] raise self.value [ 712.265109] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.265109] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] updated_port = self._update_port( [ 712.265109] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.265109] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] _ensure_no_port_binding_failure(port) [ 712.265109] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.265109] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] raise exception.PortBindingFailed(port_id=port['id']) [ 712.265109] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] nova.exception.PortBindingFailed: Binding failed for port a6637332-9313-492b-bdde-1386cdbb1191, please check neutron logs for more information. [ 712.265109] env[62914]: ERROR nova.compute.manager [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] [ 712.265109] env[62914]: DEBUG nova.compute.utils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Binding failed for port a6637332-9313-492b-bdde-1386cdbb1191, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 712.266723] env[62914]: DEBUG nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Build of instance e24c8bf2-1616-4a5a-b6ce-0c256e78b213 was re-scheduled: Binding failed for port a6637332-9313-492b-bdde-1386cdbb1191, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 712.267161] env[62914]: DEBUG nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 712.267390] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Acquiring lock "refresh_cache-e24c8bf2-1616-4a5a-b6ce-0c256e78b213" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.267533] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Acquired lock "refresh_cache-e24c8bf2-1616-4a5a-b6ce-0c256e78b213" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.267748] env[62914]: DEBUG nova.network.neutron [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 712.269176] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.435s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.274596] env[62914]: DEBUG oslo_concurrency.lockutils [None req-97e63cfe-8377-477f-b103-9bb2f69c4aea tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Lock "0ac803ee-6303-463f-91e8-8263c0219b0a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.111s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.466350] env[62914]: INFO nova.compute.manager [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] [instance: 3ce62ec7-6274-441f-989b-6f27a64b0c18] Took 1.02 seconds to deallocate network for instance. [ 712.778573] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 712.808016] env[62914]: DEBUG nova.network.neutron [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.923794] env[62914]: DEBUG nova.network.neutron [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.110935] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d8caaac-ae78-4d8c-b03a-5badbc6d6d61 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.118861] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e5edee-cfc5-45e3-90b8-039adf81d636 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.149972] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2997a770-c0f8-4e83-8bd2-80f7b581084e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.157872] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ff333d-5bcf-41e5-8cf1-c508a68116c7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.171373] env[62914]: DEBUG nova.compute.provider_tree [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.263330] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Acquiring lock "8a342650-70f0-4d5b-8086-6450dcc8d95b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.263560] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Lock "8a342650-70f0-4d5b-8086-6450dcc8d95b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.297352] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.425975] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Releasing lock "refresh_cache-e24c8bf2-1616-4a5a-b6ce-0c256e78b213" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.426351] env[62914]: DEBUG nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 713.426470] env[62914]: DEBUG nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.426678] env[62914]: DEBUG nova.network.neutron [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 713.442867] env[62914]: DEBUG nova.network.neutron [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.497309] env[62914]: INFO nova.scheduler.client.report [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Deleted allocations for instance 3ce62ec7-6274-441f-989b-6f27a64b0c18 [ 713.674413] env[62914]: DEBUG nova.scheduler.client.report [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 713.946262] env[62914]: DEBUG nova.network.neutron [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.007997] env[62914]: DEBUG oslo_concurrency.lockutils [None req-35da52c1-55ec-4703-8f9e-6fd42e58cce6 tempest-ListImageFiltersTestJSON-1168607556 tempest-ListImageFiltersTestJSON-1168607556-project-member] Lock "3ce62ec7-6274-441f-989b-6f27a64b0c18" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.713s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.179204] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.910s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.179838] env[62914]: ERROR nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 48a08124-eb4f-427e-92bb-2c75fa5649a3, please check neutron logs for more information. [ 714.179838] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Traceback (most recent call last): [ 714.179838] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.179838] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] self.driver.spawn(context, instance, image_meta, [ 714.179838] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 714.179838] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.179838] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.179838] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] vm_ref = self.build_virtual_machine(instance, [ 714.179838] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.179838] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.179838] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.180214] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] for vif in network_info: [ 714.180214] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.180214] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] return self._sync_wrapper(fn, *args, **kwargs) [ 714.180214] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.180214] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] self.wait() [ 714.180214] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.180214] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] self[:] = self._gt.wait() [ 714.180214] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.180214] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] return self._exit_event.wait() [ 714.180214] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 714.180214] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] current.throw(*self._exc) [ 714.180214] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.180214] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] result = function(*args, **kwargs) [ 714.180606] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 714.180606] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] return func(*args, **kwargs) [ 714.180606] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.180606] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] raise e [ 714.180606] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.180606] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] nwinfo = self.network_api.allocate_for_instance( [ 714.180606] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.180606] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] created_port_ids = self._update_ports_for_instance( [ 714.180606] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.180606] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] with excutils.save_and_reraise_exception(): [ 714.180606] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.180606] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] self.force_reraise() [ 714.180606] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.181080] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] raise self.value [ 714.181080] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.181080] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] updated_port = self._update_port( [ 714.181080] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.181080] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] _ensure_no_port_binding_failure(port) [ 714.181080] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.181080] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] raise exception.PortBindingFailed(port_id=port['id']) [ 714.181080] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] nova.exception.PortBindingFailed: Binding failed for port 48a08124-eb4f-427e-92bb-2c75fa5649a3, please check neutron logs for more information. [ 714.181080] env[62914]: ERROR nova.compute.manager [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] [ 714.181080] env[62914]: DEBUG nova.compute.utils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Binding failed for port 48a08124-eb4f-427e-92bb-2c75fa5649a3, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 714.182274] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.600s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.184921] env[62914]: INFO nova.compute.claims [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.190268] env[62914]: DEBUG nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Build of instance ad283bc7-b422-4836-a503-d3d9ae4fe3c1 was re-scheduled: Binding failed for port 48a08124-eb4f-427e-92bb-2c75fa5649a3, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 714.190268] env[62914]: DEBUG nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 714.190268] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Acquiring lock "refresh_cache-ad283bc7-b422-4836-a503-d3d9ae4fe3c1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.190268] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Acquired lock "refresh_cache-ad283bc7-b422-4836-a503-d3d9ae4fe3c1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.190429] env[62914]: DEBUG nova.network.neutron [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 714.449402] env[62914]: INFO nova.compute.manager [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] [instance: e24c8bf2-1616-4a5a-b6ce-0c256e78b213] Took 1.02 seconds to deallocate network for instance. [ 714.510214] env[62914]: DEBUG nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 714.710861] env[62914]: DEBUG nova.network.neutron [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.769293] env[62914]: DEBUG nova.network.neutron [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.044264] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.271833] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Releasing lock "refresh_cache-ad283bc7-b422-4836-a503-d3d9ae4fe3c1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.272090] env[62914]: DEBUG nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 715.272256] env[62914]: DEBUG nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 715.272419] env[62914]: DEBUG nova.network.neutron [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 715.300289] env[62914]: DEBUG nova.network.neutron [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.485222] env[62914]: INFO nova.scheduler.client.report [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Deleted allocations for instance e24c8bf2-1616-4a5a-b6ce-0c256e78b213 [ 715.524385] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c027992-9309-4f07-a7c7-e9e451e7717d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.532596] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77204656-1423-4030-a3f8-b8ec24604041 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.564126] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329a1496-40de-4d5c-ad19-25b55b7bc0b7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.571997] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cede0c6-78d2-4d71-9fcb-6997d87a897c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.585135] env[62914]: DEBUG nova.compute.provider_tree [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.805749] env[62914]: DEBUG nova.network.neutron [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.995884] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a697d973-eb03-497c-ad8f-708de6b7c0ea tempest-ServerTagsTestJSON-1174078800 tempest-ServerTagsTestJSON-1174078800-project-member] Lock "e24c8bf2-1616-4a5a-b6ce-0c256e78b213" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.574s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.088251] env[62914]: DEBUG nova.scheduler.client.report [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.308746] env[62914]: INFO nova.compute.manager [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] [instance: ad283bc7-b422-4836-a503-d3d9ae4fe3c1] Took 1.04 seconds to deallocate network for instance. [ 716.498083] env[62914]: DEBUG nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 716.593315] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.593836] env[62914]: DEBUG nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 716.596323] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.118s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.597830] env[62914]: INFO nova.compute.claims [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 717.022144] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.103455] env[62914]: DEBUG nova.compute.utils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 717.113742] env[62914]: DEBUG nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 717.113742] env[62914]: DEBUG nova.network.neutron [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 717.164260] env[62914]: DEBUG nova.policy [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '43ed8e8b366e4b07aff91219c5725993', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2221f1d980df4f19ae4557722ad36e92', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 717.347735] env[62914]: INFO nova.scheduler.client.report [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Deleted allocations for instance ad283bc7-b422-4836-a503-d3d9ae4fe3c1 [ 717.456306] env[62914]: DEBUG nova.network.neutron [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Successfully created port: 7b798aef-dc26-44e5-b96c-1c00a7e2fe61 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.610821] env[62914]: DEBUG nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 717.856079] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0464d57-3e21-4f1d-b65a-085426a64cf3 tempest-MigrationsAdminTest-1089847692 tempest-MigrationsAdminTest-1089847692-project-member] Lock "ad283bc7-b422-4836-a503-d3d9ae4fe3c1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.020s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.974321] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f34ea2-bd6f-4729-b50f-38c24e88612a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.985264] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d187fef3-4126-4f51-a4fc-d2e93f65855c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.015882] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc962ac9-4c6b-4257-bf29-9028dc873e9d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.023861] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0854c2-3691-469e-9f59-4a5af32e3905 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.039349] env[62914]: DEBUG nova.compute.provider_tree [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.360812] env[62914]: DEBUG nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 718.535182] env[62914]: DEBUG nova.compute.manager [req-cc9ea4df-5629-42ca-a63f-a3988585c1de req-6b1ebedf-02a1-4eaf-b143-d1a95024bc94 service nova] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Received event network-changed-7b798aef-dc26-44e5-b96c-1c00a7e2fe61 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.535182] env[62914]: DEBUG nova.compute.manager [req-cc9ea4df-5629-42ca-a63f-a3988585c1de req-6b1ebedf-02a1-4eaf-b143-d1a95024bc94 service nova] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Refreshing instance network info cache due to event network-changed-7b798aef-dc26-44e5-b96c-1c00a7e2fe61. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 718.535182] env[62914]: DEBUG oslo_concurrency.lockutils [req-cc9ea4df-5629-42ca-a63f-a3988585c1de req-6b1ebedf-02a1-4eaf-b143-d1a95024bc94 service nova] Acquiring lock "refresh_cache-9edfad51-10a5-4e0c-a337-2de909c361cb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.535182] env[62914]: DEBUG oslo_concurrency.lockutils [req-cc9ea4df-5629-42ca-a63f-a3988585c1de req-6b1ebedf-02a1-4eaf-b143-d1a95024bc94 service nova] Acquired lock "refresh_cache-9edfad51-10a5-4e0c-a337-2de909c361cb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.535182] env[62914]: DEBUG nova.network.neutron [req-cc9ea4df-5629-42ca-a63f-a3988585c1de req-6b1ebedf-02a1-4eaf-b143-d1a95024bc94 service nova] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Refreshing network info cache for port 7b798aef-dc26-44e5-b96c-1c00a7e2fe61 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 718.543093] env[62914]: DEBUG nova.scheduler.client.report [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.622102] env[62914]: DEBUG nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 718.653808] env[62914]: DEBUG nova.virt.hardware [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 718.654135] env[62914]: DEBUG nova.virt.hardware [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 718.654279] env[62914]: DEBUG nova.virt.hardware [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.654432] env[62914]: DEBUG nova.virt.hardware [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 718.654573] env[62914]: DEBUG nova.virt.hardware [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.654765] env[62914]: DEBUG nova.virt.hardware [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 718.654911] env[62914]: DEBUG nova.virt.hardware [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 718.656016] env[62914]: DEBUG nova.virt.hardware [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 718.656016] env[62914]: DEBUG nova.virt.hardware [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 718.656016] env[62914]: DEBUG nova.virt.hardware [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 718.656016] env[62914]: DEBUG nova.virt.hardware [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.656523] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5151ea11-49f8-4e9b-a3de-538df4e58c1d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.664717] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14874a94-b7d0-4b46-ab67-968318c742ff {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.886829] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.888689] env[62914]: ERROR nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7b798aef-dc26-44e5-b96c-1c00a7e2fe61, please check neutron logs for more information. [ 718.888689] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 718.888689] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.888689] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 718.888689] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.888689] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 718.888689] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.888689] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 718.888689] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.888689] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 718.888689] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.888689] env[62914]: ERROR nova.compute.manager raise self.value [ 718.888689] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.888689] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 718.888689] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.888689] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 718.889243] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.889243] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 718.889243] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7b798aef-dc26-44e5-b96c-1c00a7e2fe61, please check neutron logs for more information. [ 718.889243] env[62914]: ERROR nova.compute.manager [ 718.889243] env[62914]: Traceback (most recent call last): [ 718.889243] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 718.889243] env[62914]: listener.cb(fileno) [ 718.889243] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.889243] env[62914]: result = function(*args, **kwargs) [ 718.889243] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.889243] env[62914]: return func(*args, **kwargs) [ 718.889243] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.889243] env[62914]: raise e [ 718.889243] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.889243] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 718.889243] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.889243] env[62914]: created_port_ids = self._update_ports_for_instance( [ 718.889243] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.889243] env[62914]: with excutils.save_and_reraise_exception(): [ 718.889243] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.889243] env[62914]: self.force_reraise() [ 718.889243] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.889243] env[62914]: raise self.value [ 718.889243] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.889243] env[62914]: updated_port = self._update_port( [ 718.889243] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.889243] env[62914]: _ensure_no_port_binding_failure(port) [ 718.889243] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.889243] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 718.890335] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 7b798aef-dc26-44e5-b96c-1c00a7e2fe61, please check neutron logs for more information. [ 718.890335] env[62914]: Removing descriptor: 19 [ 718.890335] env[62914]: ERROR nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7b798aef-dc26-44e5-b96c-1c00a7e2fe61, please check neutron logs for more information. [ 718.890335] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Traceback (most recent call last): [ 718.890335] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 718.890335] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] yield resources [ 718.890335] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 718.890335] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] self.driver.spawn(context, instance, image_meta, [ 718.890335] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 718.890335] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.890335] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.890335] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] vm_ref = self.build_virtual_machine(instance, [ 718.890734] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.890734] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.890734] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.890734] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] for vif in network_info: [ 718.890734] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.890734] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] return self._sync_wrapper(fn, *args, **kwargs) [ 718.890734] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.890734] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] self.wait() [ 718.890734] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.890734] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] self[:] = self._gt.wait() [ 718.890734] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.890734] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] return self._exit_event.wait() [ 718.890734] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 718.891150] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] result = hub.switch() [ 718.891150] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 718.891150] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] return self.greenlet.switch() [ 718.891150] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.891150] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] result = function(*args, **kwargs) [ 718.891150] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.891150] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] return func(*args, **kwargs) [ 718.891150] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.891150] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] raise e [ 718.891150] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.891150] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] nwinfo = self.network_api.allocate_for_instance( [ 718.891150] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.891150] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] created_port_ids = self._update_ports_for_instance( [ 718.892136] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.892136] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] with excutils.save_and_reraise_exception(): [ 718.892136] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.892136] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] self.force_reraise() [ 718.892136] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.892136] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] raise self.value [ 718.892136] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.892136] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] updated_port = self._update_port( [ 718.892136] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.892136] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] _ensure_no_port_binding_failure(port) [ 718.892136] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.892136] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] raise exception.PortBindingFailed(port_id=port['id']) [ 718.892539] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] nova.exception.PortBindingFailed: Binding failed for port 7b798aef-dc26-44e5-b96c-1c00a7e2fe61, please check neutron logs for more information. [ 718.892539] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] [ 718.892539] env[62914]: INFO nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Terminating instance [ 718.893084] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Acquiring lock "refresh_cache-9edfad51-10a5-4e0c-a337-2de909c361cb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.050139] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.453s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.052112] env[62914]: DEBUG nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 719.054032] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.560s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.188784] env[62914]: DEBUG nova.network.neutron [req-cc9ea4df-5629-42ca-a63f-a3988585c1de req-6b1ebedf-02a1-4eaf-b143-d1a95024bc94 service nova] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.337542] env[62914]: DEBUG nova.network.neutron [req-cc9ea4df-5629-42ca-a63f-a3988585c1de req-6b1ebedf-02a1-4eaf-b143-d1a95024bc94 service nova] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.557680] env[62914]: DEBUG nova.compute.utils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 719.564021] env[62914]: DEBUG nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 719.564021] env[62914]: DEBUG nova.network.neutron [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 719.630912] env[62914]: DEBUG nova.policy [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '43ed8e8b366e4b07aff91219c5725993', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2221f1d980df4f19ae4557722ad36e92', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 719.843301] env[62914]: DEBUG oslo_concurrency.lockutils [req-cc9ea4df-5629-42ca-a63f-a3988585c1de req-6b1ebedf-02a1-4eaf-b143-d1a95024bc94 service nova] Releasing lock "refresh_cache-9edfad51-10a5-4e0c-a337-2de909c361cb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.844128] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Acquired lock "refresh_cache-9edfad51-10a5-4e0c-a337-2de909c361cb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.844313] env[62914]: DEBUG nova.network.neutron [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.939358] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f1da42-5135-40b0-9c2c-034240082784 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.948069] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e759962-d1d3-4091-a249-00df8b7e92fc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.985290] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339aea5c-59cb-4d09-b196-d40fd8ae6ef8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.998018] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b1632b-eab0-4a5e-a80a-d39f9e344022 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.014481] env[62914]: DEBUG nova.compute.provider_tree [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.064429] env[62914]: DEBUG nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 720.118758] env[62914]: DEBUG nova.network.neutron [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Successfully created port: 1195773f-b43f-4e3a-944c-f6116a5b8ac7 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 720.364683] env[62914]: DEBUG nova.network.neutron [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.460135] env[62914]: DEBUG nova.network.neutron [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.520097] env[62914]: DEBUG nova.scheduler.client.report [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.528582] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquiring lock "87ef6b06-f699-4de3-8b89-854717074406" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.530609] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Lock "87ef6b06-f699-4de3-8b89-854717074406" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.638224] env[62914]: DEBUG nova.compute.manager [req-4c0a098b-d4ff-493f-8ddb-744a2a104e46 req-08c2141a-5b3d-4953-a551-362cc0c116bb service nova] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Received event network-vif-deleted-7b798aef-dc26-44e5-b96c-1c00a7e2fe61 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.962578] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Releasing lock "refresh_cache-9edfad51-10a5-4e0c-a337-2de909c361cb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.963124] env[62914]: DEBUG nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 720.963390] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 720.963608] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d61b7bc5-d8fd-482b-9f5e-b40a6a78284c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.975065] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318f25c0-af88-4476-805c-f0449df5a86d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.001505] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9edfad51-10a5-4e0c-a337-2de909c361cb could not be found. [ 721.001772] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 721.002971] env[62914]: INFO nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 721.002971] env[62914]: DEBUG oslo.service.loopingcall [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 721.002971] env[62914]: DEBUG nova.compute.manager [-] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 721.002971] env[62914]: DEBUG nova.network.neutron [-] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 721.017484] env[62914]: DEBUG nova.network.neutron [-] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.032331] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.979s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.033259] env[62914]: ERROR nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7bcf9042-0227-4463-ae3a-769768b437a5, please check neutron logs for more information. [ 721.033259] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Traceback (most recent call last): [ 721.033259] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.033259] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] self.driver.spawn(context, instance, image_meta, [ 721.033259] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 721.033259] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.033259] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.033259] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] vm_ref = self.build_virtual_machine(instance, [ 721.033259] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.033259] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.033259] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.033832] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] for vif in network_info: [ 721.033832] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 721.033832] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] return self._sync_wrapper(fn, *args, **kwargs) [ 721.033832] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 721.033832] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] self.wait() [ 721.033832] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 721.033832] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] self[:] = self._gt.wait() [ 721.033832] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.033832] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] return self._exit_event.wait() [ 721.033832] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 721.033832] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] result = hub.switch() [ 721.033832] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 721.033832] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] return self.greenlet.switch() [ 721.034451] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.034451] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] result = function(*args, **kwargs) [ 721.034451] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 721.034451] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] return func(*args, **kwargs) [ 721.034451] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.034451] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] raise e [ 721.034451] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.034451] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] nwinfo = self.network_api.allocate_for_instance( [ 721.034451] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.034451] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] created_port_ids = self._update_ports_for_instance( [ 721.034451] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.034451] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] with excutils.save_and_reraise_exception(): [ 721.034451] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.034977] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] self.force_reraise() [ 721.034977] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.034977] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] raise self.value [ 721.034977] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.034977] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] updated_port = self._update_port( [ 721.034977] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.034977] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] _ensure_no_port_binding_failure(port) [ 721.034977] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.034977] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] raise exception.PortBindingFailed(port_id=port['id']) [ 721.034977] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] nova.exception.PortBindingFailed: Binding failed for port 7bcf9042-0227-4463-ae3a-769768b437a5, please check neutron logs for more information. [ 721.034977] env[62914]: ERROR nova.compute.manager [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] [ 721.035351] env[62914]: DEBUG nova.compute.utils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Binding failed for port 7bcf9042-0227-4463-ae3a-769768b437a5, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 721.036266] env[62914]: DEBUG nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Build of instance 67554dc7-fb72-47e1-8201-f5fda035a2d3 was re-scheduled: Binding failed for port 7bcf9042-0227-4463-ae3a-769768b437a5, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 721.036881] env[62914]: DEBUG nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 721.037191] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Acquiring lock "refresh_cache-67554dc7-fb72-47e1-8201-f5fda035a2d3" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.037390] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Acquired lock "refresh_cache-67554dc7-fb72-47e1-8201-f5fda035a2d3" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.037605] env[62914]: DEBUG nova.network.neutron [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.042020] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.163s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.042020] env[62914]: INFO nova.compute.claims [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.075251] env[62914]: DEBUG nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 721.102673] env[62914]: DEBUG nova.virt.hardware [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 721.102918] env[62914]: DEBUG nova.virt.hardware [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 721.103191] env[62914]: DEBUG nova.virt.hardware [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 721.103267] env[62914]: DEBUG nova.virt.hardware [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 721.103397] env[62914]: DEBUG nova.virt.hardware [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 721.103534] env[62914]: DEBUG nova.virt.hardware [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 721.103733] env[62914]: DEBUG nova.virt.hardware [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 721.103877] env[62914]: DEBUG nova.virt.hardware [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 721.104476] env[62914]: DEBUG nova.virt.hardware [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 721.104687] env[62914]: DEBUG nova.virt.hardware [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 721.104866] env[62914]: DEBUG nova.virt.hardware [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 721.105977] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ba70b1-885c-4df3-8bf2-31dbce43a4cc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.115942] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6536152-9c9a-46b6-afc1-586e720741ae {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.519836] env[62914]: DEBUG nova.network.neutron [-] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.553182] env[62914]: ERROR nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1195773f-b43f-4e3a-944c-f6116a5b8ac7, please check neutron logs for more information. [ 721.553182] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 721.553182] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.553182] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 721.553182] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.553182] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 721.553182] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.553182] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 721.553182] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.553182] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 721.553182] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.553182] env[62914]: ERROR nova.compute.manager raise self.value [ 721.553182] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.553182] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 721.553182] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.553182] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 721.553787] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.553787] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 721.553787] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1195773f-b43f-4e3a-944c-f6116a5b8ac7, please check neutron logs for more information. [ 721.553787] env[62914]: ERROR nova.compute.manager [ 721.553787] env[62914]: Traceback (most recent call last): [ 721.553787] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 721.553787] env[62914]: listener.cb(fileno) [ 721.553787] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.553787] env[62914]: result = function(*args, **kwargs) [ 721.553787] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 721.553787] env[62914]: return func(*args, **kwargs) [ 721.553787] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.553787] env[62914]: raise e [ 721.553787] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.553787] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 721.553787] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.553787] env[62914]: created_port_ids = self._update_ports_for_instance( [ 721.553787] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.553787] env[62914]: with excutils.save_and_reraise_exception(): [ 721.553787] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.553787] env[62914]: self.force_reraise() [ 721.553787] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.553787] env[62914]: raise self.value [ 721.553787] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.553787] env[62914]: updated_port = self._update_port( [ 721.553787] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.553787] env[62914]: _ensure_no_port_binding_failure(port) [ 721.553787] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.553787] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 721.554759] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 1195773f-b43f-4e3a-944c-f6116a5b8ac7, please check neutron logs for more information. [ 721.554759] env[62914]: Removing descriptor: 19 [ 721.554759] env[62914]: ERROR nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1195773f-b43f-4e3a-944c-f6116a5b8ac7, please check neutron logs for more information. [ 721.554759] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Traceback (most recent call last): [ 721.554759] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 721.554759] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] yield resources [ 721.554759] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.554759] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] self.driver.spawn(context, instance, image_meta, [ 721.554759] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 721.554759] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.554759] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.554759] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] vm_ref = self.build_virtual_machine(instance, [ 721.555182] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.555182] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.555182] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.555182] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] for vif in network_info: [ 721.555182] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 721.555182] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] return self._sync_wrapper(fn, *args, **kwargs) [ 721.555182] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 721.555182] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] self.wait() [ 721.555182] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 721.555182] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] self[:] = self._gt.wait() [ 721.555182] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.555182] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] return self._exit_event.wait() [ 721.555182] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 721.555612] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] result = hub.switch() [ 721.555612] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 721.555612] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] return self.greenlet.switch() [ 721.555612] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.555612] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] result = function(*args, **kwargs) [ 721.555612] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 721.555612] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] return func(*args, **kwargs) [ 721.555612] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.555612] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] raise e [ 721.555612] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.555612] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] nwinfo = self.network_api.allocate_for_instance( [ 721.555612] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.555612] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] created_port_ids = self._update_ports_for_instance( [ 721.556297] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.556297] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] with excutils.save_and_reraise_exception(): [ 721.556297] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.556297] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] self.force_reraise() [ 721.556297] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.556297] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] raise self.value [ 721.556297] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.556297] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] updated_port = self._update_port( [ 721.556297] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.556297] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] _ensure_no_port_binding_failure(port) [ 721.556297] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.556297] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] raise exception.PortBindingFailed(port_id=port['id']) [ 721.556671] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] nova.exception.PortBindingFailed: Binding failed for port 1195773f-b43f-4e3a-944c-f6116a5b8ac7, please check neutron logs for more information. [ 721.556671] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] [ 721.556671] env[62914]: INFO nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Terminating instance [ 721.556671] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Acquiring lock "refresh_cache-e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.556671] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Acquired lock "refresh_cache-e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.556886] env[62914]: DEBUG nova.network.neutron [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.576733] env[62914]: DEBUG nova.network.neutron [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.714369] env[62914]: DEBUG nova.network.neutron [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.025111] env[62914]: INFO nova.compute.manager [-] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Took 1.02 seconds to deallocate network for instance. [ 722.027501] env[62914]: DEBUG nova.compute.claims [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 722.027671] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.066732] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Acquiring lock "9fab500c-344c-46cd-b060-c8f1aa0f1cba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.066732] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Lock "9fab500c-344c-46cd-b060-c8f1aa0f1cba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.077061] env[62914]: DEBUG nova.network.neutron [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.167039] env[62914]: DEBUG nova.network.neutron [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.216566] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Releasing lock "refresh_cache-67554dc7-fb72-47e1-8201-f5fda035a2d3" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.216566] env[62914]: DEBUG nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 722.216677] env[62914]: DEBUG nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.216858] env[62914]: DEBUG nova.network.neutron [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.231849] env[62914]: DEBUG nova.network.neutron [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.358208] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c8c0bf-26c1-4dd6-80cc-10f72b78b047 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.365825] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf79729f-77bd-4164-ba99-6e71bd6f2cd9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.397423] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac2ea8e0-9145-444a-9f80-db7e20ce6a57 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.404664] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f3c035-c05d-4474-ab1e-adbf0dee1ad3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.418699] env[62914]: DEBUG nova.compute.provider_tree [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.670172] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Releasing lock "refresh_cache-e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.670609] env[62914]: DEBUG nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 722.670800] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 722.671119] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64ecedb0-8556-49cf-af41-df73bdeaaaed {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.681410] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55bfcbb-5eb3-497f-abff-b3bf11209460 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.696094] env[62914]: DEBUG nova.compute.manager [req-c5adc8f9-82f4-428b-b54c-0595bc2800e1 req-154840a8-5e60-4e62-abc4-8c9d470c32e2 service nova] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Received event network-changed-1195773f-b43f-4e3a-944c-f6116a5b8ac7 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.696094] env[62914]: DEBUG nova.compute.manager [req-c5adc8f9-82f4-428b-b54c-0595bc2800e1 req-154840a8-5e60-4e62-abc4-8c9d470c32e2 service nova] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Refreshing instance network info cache due to event network-changed-1195773f-b43f-4e3a-944c-f6116a5b8ac7. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 722.696185] env[62914]: DEBUG oslo_concurrency.lockutils [req-c5adc8f9-82f4-428b-b54c-0595bc2800e1 req-154840a8-5e60-4e62-abc4-8c9d470c32e2 service nova] Acquiring lock "refresh_cache-e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.696243] env[62914]: DEBUG oslo_concurrency.lockutils [req-c5adc8f9-82f4-428b-b54c-0595bc2800e1 req-154840a8-5e60-4e62-abc4-8c9d470c32e2 service nova] Acquired lock "refresh_cache-e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.696392] env[62914]: DEBUG nova.network.neutron [req-c5adc8f9-82f4-428b-b54c-0595bc2800e1 req-154840a8-5e60-4e62-abc4-8c9d470c32e2 service nova] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Refreshing network info cache for port 1195773f-b43f-4e3a-944c-f6116a5b8ac7 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 722.709608] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6 could not be found. [ 722.709781] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 722.710055] env[62914]: INFO nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 722.710308] env[62914]: DEBUG oslo.service.loopingcall [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.710516] env[62914]: DEBUG nova.compute.manager [-] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.710677] env[62914]: DEBUG nova.network.neutron [-] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.726421] env[62914]: DEBUG nova.network.neutron [-] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.735342] env[62914]: DEBUG nova.network.neutron [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.922081] env[62914]: DEBUG nova.scheduler.client.report [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.214810] env[62914]: DEBUG nova.network.neutron [req-c5adc8f9-82f4-428b-b54c-0595bc2800e1 req-154840a8-5e60-4e62-abc4-8c9d470c32e2 service nova] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.228797] env[62914]: DEBUG nova.network.neutron [-] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.237878] env[62914]: INFO nova.compute.manager [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] [instance: 67554dc7-fb72-47e1-8201-f5fda035a2d3] Took 1.02 seconds to deallocate network for instance. [ 723.299798] env[62914]: DEBUG nova.network.neutron [req-c5adc8f9-82f4-428b-b54c-0595bc2800e1 req-154840a8-5e60-4e62-abc4-8c9d470c32e2 service nova] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.426522] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.387s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.427164] env[62914]: DEBUG nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 723.429887] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.820s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.731509] env[62914]: INFO nova.compute.manager [-] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Took 1.02 seconds to deallocate network for instance. [ 723.734267] env[62914]: DEBUG nova.compute.claims [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 723.734446] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.802501] env[62914]: DEBUG oslo_concurrency.lockutils [req-c5adc8f9-82f4-428b-b54c-0595bc2800e1 req-154840a8-5e60-4e62-abc4-8c9d470c32e2 service nova] Releasing lock "refresh_cache-e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.802839] env[62914]: DEBUG nova.compute.manager [req-c5adc8f9-82f4-428b-b54c-0595bc2800e1 req-154840a8-5e60-4e62-abc4-8c9d470c32e2 service nova] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Received event network-vif-deleted-1195773f-b43f-4e3a-944c-f6116a5b8ac7 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 723.935120] env[62914]: DEBUG nova.compute.utils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 723.940495] env[62914]: DEBUG nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 723.940495] env[62914]: DEBUG nova.network.neutron [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 723.986119] env[62914]: DEBUG nova.policy [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '82a857722ff043c994009b16db6373aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f897e1c819f44362857de4d4398ca613', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 724.271278] env[62914]: INFO nova.scheduler.client.report [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Deleted allocations for instance 67554dc7-fb72-47e1-8201-f5fda035a2d3 [ 724.284019] env[62914]: DEBUG nova.network.neutron [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Successfully created port: 1ac3a328-5cd6-4343-8358-b1ea3e945ea2 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.335112] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4ffda2-44f0-455a-b5ff-a2f6832e4e31 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.341702] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1177160e-db76-4041-9e1c-9aaada1fc6cf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.375830] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0b7f0e-b7f4-44e4-b70a-564d8fc7bbeb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.385880] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5d4611-0a8b-4c65-aff8-ee23dca834ea {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.402446] env[62914]: DEBUG nova.compute.provider_tree [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.442861] env[62914]: DEBUG nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 724.783025] env[62914]: DEBUG oslo_concurrency.lockutils [None req-77b419ec-ff05-490b-9b3e-5a839243cab2 tempest-FloatingIPsAssociationTestJSON-344624329 tempest-FloatingIPsAssociationTestJSON-344624329-project-member] Lock "67554dc7-fb72-47e1-8201-f5fda035a2d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.217s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.906832] env[62914]: DEBUG nova.scheduler.client.report [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.220932] env[62914]: DEBUG nova.compute.manager [req-520ed503-5b14-4825-ba4a-0bdfda247884 req-04db7808-a4c5-45ad-b8fe-f9ffa21b1d88 service nova] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Received event network-changed-1ac3a328-5cd6-4343-8358-b1ea3e945ea2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.221139] env[62914]: DEBUG nova.compute.manager [req-520ed503-5b14-4825-ba4a-0bdfda247884 req-04db7808-a4c5-45ad-b8fe-f9ffa21b1d88 service nova] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Refreshing instance network info cache due to event network-changed-1ac3a328-5cd6-4343-8358-b1ea3e945ea2. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 725.221351] env[62914]: DEBUG oslo_concurrency.lockutils [req-520ed503-5b14-4825-ba4a-0bdfda247884 req-04db7808-a4c5-45ad-b8fe-f9ffa21b1d88 service nova] Acquiring lock "refresh_cache-8399d066-43bc-4766-97c8-89d74f3283f0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.221489] env[62914]: DEBUG oslo_concurrency.lockutils [req-520ed503-5b14-4825-ba4a-0bdfda247884 req-04db7808-a4c5-45ad-b8fe-f9ffa21b1d88 service nova] Acquired lock "refresh_cache-8399d066-43bc-4766-97c8-89d74f3283f0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.221642] env[62914]: DEBUG nova.network.neutron [req-520ed503-5b14-4825-ba4a-0bdfda247884 req-04db7808-a4c5-45ad-b8fe-f9ffa21b1d88 service nova] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Refreshing network info cache for port 1ac3a328-5cd6-4343-8358-b1ea3e945ea2 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 725.287407] env[62914]: DEBUG nova.compute.manager [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 725.414570] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.985s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.415289] env[62914]: ERROR nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e5560f64-e00b-4658-87c5-b88d50d90135, please check neutron logs for more information. [ 725.415289] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Traceback (most recent call last): [ 725.415289] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.415289] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] self.driver.spawn(context, instance, image_meta, [ 725.415289] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 725.415289] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.415289] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.415289] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] vm_ref = self.build_virtual_machine(instance, [ 725.415289] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.415289] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.415289] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.415633] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] for vif in network_info: [ 725.415633] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 725.415633] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] return self._sync_wrapper(fn, *args, **kwargs) [ 725.415633] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 725.415633] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] self.wait() [ 725.415633] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 725.415633] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] self[:] = self._gt.wait() [ 725.415633] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.415633] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] return self._exit_event.wait() [ 725.415633] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 725.415633] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] current.throw(*self._exc) [ 725.415633] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.415633] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] result = function(*args, **kwargs) [ 725.415993] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 725.415993] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] return func(*args, **kwargs) [ 725.415993] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.415993] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] raise e [ 725.415993] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.415993] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] nwinfo = self.network_api.allocate_for_instance( [ 725.415993] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.415993] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] created_port_ids = self._update_ports_for_instance( [ 725.415993] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.415993] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] with excutils.save_and_reraise_exception(): [ 725.415993] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.415993] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] self.force_reraise() [ 725.415993] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.416358] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] raise self.value [ 725.416358] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.416358] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] updated_port = self._update_port( [ 725.416358] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.416358] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] _ensure_no_port_binding_failure(port) [ 725.416358] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.416358] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] raise exception.PortBindingFailed(port_id=port['id']) [ 725.416358] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] nova.exception.PortBindingFailed: Binding failed for port e5560f64-e00b-4658-87c5-b88d50d90135, please check neutron logs for more information. [ 725.416358] env[62914]: ERROR nova.compute.manager [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] [ 725.416358] env[62914]: DEBUG nova.compute.utils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Binding failed for port e5560f64-e00b-4658-87c5-b88d50d90135, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 725.417757] env[62914]: DEBUG nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Build of instance 821366f1-f1b0-4b78-8f7e-db8fd3bb123e was re-scheduled: Binding failed for port e5560f64-e00b-4658-87c5-b88d50d90135, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 725.418741] env[62914]: DEBUG nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 725.418741] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Acquiring lock "refresh_cache-821366f1-f1b0-4b78-8f7e-db8fd3bb123e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.418741] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Acquired lock "refresh_cache-821366f1-f1b0-4b78-8f7e-db8fd3bb123e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.418741] env[62914]: DEBUG nova.network.neutron [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.420126] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.164s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.421679] env[62914]: INFO nova.compute.claims [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 725.435276] env[62914]: ERROR nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1ac3a328-5cd6-4343-8358-b1ea3e945ea2, please check neutron logs for more information. [ 725.435276] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 725.435276] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.435276] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 725.435276] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.435276] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 725.435276] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.435276] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 725.435276] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.435276] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 725.435276] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.435276] env[62914]: ERROR nova.compute.manager raise self.value [ 725.435276] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.435276] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 725.435276] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.435276] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 725.435746] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.435746] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 725.435746] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1ac3a328-5cd6-4343-8358-b1ea3e945ea2, please check neutron logs for more information. [ 725.435746] env[62914]: ERROR nova.compute.manager [ 725.435746] env[62914]: Traceback (most recent call last): [ 725.435746] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 725.435746] env[62914]: listener.cb(fileno) [ 725.435746] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.435746] env[62914]: result = function(*args, **kwargs) [ 725.435746] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 725.435746] env[62914]: return func(*args, **kwargs) [ 725.435746] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.435746] env[62914]: raise e [ 725.435746] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.435746] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 725.435746] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.435746] env[62914]: created_port_ids = self._update_ports_for_instance( [ 725.435746] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.435746] env[62914]: with excutils.save_and_reraise_exception(): [ 725.435746] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.435746] env[62914]: self.force_reraise() [ 725.435746] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.435746] env[62914]: raise self.value [ 725.435746] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.435746] env[62914]: updated_port = self._update_port( [ 725.435746] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.435746] env[62914]: _ensure_no_port_binding_failure(port) [ 725.435746] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.435746] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 725.436793] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 1ac3a328-5cd6-4343-8358-b1ea3e945ea2, please check neutron logs for more information. [ 725.436793] env[62914]: Removing descriptor: 15 [ 725.458113] env[62914]: DEBUG nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 725.489693] env[62914]: DEBUG nova.virt.hardware [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.489693] env[62914]: DEBUG nova.virt.hardware [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.489886] env[62914]: DEBUG nova.virt.hardware [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.490021] env[62914]: DEBUG nova.virt.hardware [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.490171] env[62914]: DEBUG nova.virt.hardware [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.490317] env[62914]: DEBUG nova.virt.hardware [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.490516] env[62914]: DEBUG nova.virt.hardware [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.490666] env[62914]: DEBUG nova.virt.hardware [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.490827] env[62914]: DEBUG nova.virt.hardware [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.490983] env[62914]: DEBUG nova.virt.hardware [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.491306] env[62914]: DEBUG nova.virt.hardware [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.495736] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853b8f0e-ff40-4481-9a0b-f09bfe2804ad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.501979] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb43dbe5-bb59-4671-b8d3-d117147c4783 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.518131] env[62914]: ERROR nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1ac3a328-5cd6-4343-8358-b1ea3e945ea2, please check neutron logs for more information. [ 725.518131] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Traceback (most recent call last): [ 725.518131] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 725.518131] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] yield resources [ 725.518131] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.518131] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] self.driver.spawn(context, instance, image_meta, [ 725.518131] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 725.518131] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.518131] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.518131] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] vm_ref = self.build_virtual_machine(instance, [ 725.518131] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.518519] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.518519] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.518519] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] for vif in network_info: [ 725.518519] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 725.518519] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] return self._sync_wrapper(fn, *args, **kwargs) [ 725.518519] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 725.518519] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] self.wait() [ 725.518519] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 725.518519] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] self[:] = self._gt.wait() [ 725.518519] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.518519] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] return self._exit_event.wait() [ 725.518519] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 725.518519] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] current.throw(*self._exc) [ 725.518870] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.518870] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] result = function(*args, **kwargs) [ 725.518870] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 725.518870] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] return func(*args, **kwargs) [ 725.518870] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.518870] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] raise e [ 725.518870] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.518870] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] nwinfo = self.network_api.allocate_for_instance( [ 725.518870] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.518870] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] created_port_ids = self._update_ports_for_instance( [ 725.518870] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.518870] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] with excutils.save_and_reraise_exception(): [ 725.518870] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.519256] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] self.force_reraise() [ 725.519256] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.519256] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] raise self.value [ 725.519256] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.519256] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] updated_port = self._update_port( [ 725.519256] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.519256] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] _ensure_no_port_binding_failure(port) [ 725.519256] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.519256] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] raise exception.PortBindingFailed(port_id=port['id']) [ 725.519256] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] nova.exception.PortBindingFailed: Binding failed for port 1ac3a328-5cd6-4343-8358-b1ea3e945ea2, please check neutron logs for more information. [ 725.519256] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] [ 725.519256] env[62914]: INFO nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Terminating instance [ 725.521565] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "refresh_cache-8399d066-43bc-4766-97c8-89d74f3283f0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.763687] env[62914]: DEBUG nova.network.neutron [req-520ed503-5b14-4825-ba4a-0bdfda247884 req-04db7808-a4c5-45ad-b8fe-f9ffa21b1d88 service nova] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.812475] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.931603] env[62914]: DEBUG nova.network.neutron [req-520ed503-5b14-4825-ba4a-0bdfda247884 req-04db7808-a4c5-45ad-b8fe-f9ffa21b1d88 service nova] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.945313] env[62914]: DEBUG nova.network.neutron [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.039490] env[62914]: DEBUG nova.network.neutron [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.434591] env[62914]: DEBUG oslo_concurrency.lockutils [req-520ed503-5b14-4825-ba4a-0bdfda247884 req-04db7808-a4c5-45ad-b8fe-f9ffa21b1d88 service nova] Releasing lock "refresh_cache-8399d066-43bc-4766-97c8-89d74f3283f0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.435029] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquired lock "refresh_cache-8399d066-43bc-4766-97c8-89d74f3283f0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.435226] env[62914]: DEBUG nova.network.neutron [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 726.544809] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Releasing lock "refresh_cache-821366f1-f1b0-4b78-8f7e-db8fd3bb123e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.544809] env[62914]: DEBUG nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 726.544809] env[62914]: DEBUG nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.544809] env[62914]: DEBUG nova.network.neutron [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 726.566381] env[62914]: DEBUG nova.network.neutron [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.744178] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.744392] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.804512] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61771042-b638-489a-bdcb-e14fb05bce19 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.812956] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3b3a50-3cba-4025-b000-cf95d1a1e6ef {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.844978] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17eefde5-e1a0-4a47-87e4-978325edd7cc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.852526] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca029fd8-00ed-406f-9cd0-691814a022f6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.865977] env[62914]: DEBUG nova.compute.provider_tree [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.967694] env[62914]: DEBUG nova.network.neutron [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.073238] env[62914]: DEBUG nova.network.neutron [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.089870] env[62914]: DEBUG nova.network.neutron [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.273476] env[62914]: DEBUG nova.compute.manager [req-8ffd9836-365a-4607-9ec6-1a30dbbf6c19 req-5c414b3a-c6e4-4089-91c0-b5ef3c85ecb4 service nova] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Received event network-vif-deleted-1ac3a328-5cd6-4343-8358-b1ea3e945ea2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.369596] env[62914]: DEBUG nova.scheduler.client.report [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.572634] env[62914]: INFO nova.compute.manager [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] [instance: 821366f1-f1b0-4b78-8f7e-db8fd3bb123e] Took 1.03 seconds to deallocate network for instance. [ 727.594292] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Releasing lock "refresh_cache-8399d066-43bc-4766-97c8-89d74f3283f0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.594703] env[62914]: DEBUG nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 727.594927] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 727.595232] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-286ef070-0bc1-4099-bc47-27ab4cc31d8e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.604797] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a694cf-074b-4771-8bdc-2efcee392179 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.628150] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8399d066-43bc-4766-97c8-89d74f3283f0 could not be found. [ 727.628659] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 727.628659] env[62914]: INFO nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 727.628927] env[62914]: DEBUG oslo.service.loopingcall [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.629242] env[62914]: DEBUG nova.compute.manager [-] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.629361] env[62914]: DEBUG nova.network.neutron [-] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 727.644960] env[62914]: DEBUG nova.network.neutron [-] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.876820] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.457s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.879110] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 727.880801] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.584s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.882451] env[62914]: INFO nova.compute.claims [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.147698] env[62914]: DEBUG nova.network.neutron [-] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.382486] env[62914]: DEBUG nova.compute.utils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 728.383863] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 728.384044] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 728.461643] env[62914]: DEBUG nova.policy [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd9f93f76e6db40a3924a5429ab2a23f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b6b4c54c48514783a044a75dcdea02b8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 728.608235] env[62914]: INFO nova.scheduler.client.report [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Deleted allocations for instance 821366f1-f1b0-4b78-8f7e-db8fd3bb123e [ 728.650749] env[62914]: INFO nova.compute.manager [-] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Took 1.02 seconds to deallocate network for instance. [ 728.653448] env[62914]: DEBUG nova.compute.claims [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 728.654344] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.846525] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Successfully created port: ebff40d5-c9d8-4560-91fd-2d82ee0fa443 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 728.887380] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 729.119764] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0a88d917-31df-4db3-8816-1607ed4c4ce2 tempest-ServersV294TestFqdnHostnames-647685041 tempest-ServersV294TestFqdnHostnames-647685041-project-member] Lock "821366f1-f1b0-4b78-8f7e-db8fd3bb123e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.361s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.219930] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a6df11-f4e9-4f63-84d5-a0fde1987ec6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.229541] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f8f756-1790-470e-8c57-c42202f24a30 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.263880] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b265b1e4-6152-4f30-bd92-9ae8695a1a5e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.272791] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ad57b5-e195-4bfc-ba76-c2780e1fcd0a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.287016] env[62914]: DEBUG nova.compute.provider_tree [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.623740] env[62914]: DEBUG nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 729.669364] env[62914]: DEBUG nova.compute.manager [req-57badbee-32fe-4137-8a42-d4234a0f8163 req-7fa73080-dbfe-402c-914c-f7f13a02c6f7 service nova] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Received event network-changed-ebff40d5-c9d8-4560-91fd-2d82ee0fa443 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.669364] env[62914]: DEBUG nova.compute.manager [req-57badbee-32fe-4137-8a42-d4234a0f8163 req-7fa73080-dbfe-402c-914c-f7f13a02c6f7 service nova] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Refreshing instance network info cache due to event network-changed-ebff40d5-c9d8-4560-91fd-2d82ee0fa443. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 729.669364] env[62914]: DEBUG oslo_concurrency.lockutils [req-57badbee-32fe-4137-8a42-d4234a0f8163 req-7fa73080-dbfe-402c-914c-f7f13a02c6f7 service nova] Acquiring lock "refresh_cache-4c1016c4-2da8-4f4e-9497-4d680fb74535" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.669364] env[62914]: DEBUG oslo_concurrency.lockutils [req-57badbee-32fe-4137-8a42-d4234a0f8163 req-7fa73080-dbfe-402c-914c-f7f13a02c6f7 service nova] Acquired lock "refresh_cache-4c1016c4-2da8-4f4e-9497-4d680fb74535" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.669364] env[62914]: DEBUG nova.network.neutron [req-57badbee-32fe-4137-8a42-d4234a0f8163 req-7fa73080-dbfe-402c-914c-f7f13a02c6f7 service nova] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Refreshing network info cache for port ebff40d5-c9d8-4560-91fd-2d82ee0fa443 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 729.789582] env[62914]: DEBUG nova.scheduler.client.report [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.868553] env[62914]: ERROR nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ebff40d5-c9d8-4560-91fd-2d82ee0fa443, please check neutron logs for more information. [ 729.868553] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 729.868553] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.868553] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 729.868553] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.868553] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 729.868553] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.868553] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 729.868553] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.868553] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 729.868553] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.868553] env[62914]: ERROR nova.compute.manager raise self.value [ 729.868553] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.868553] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 729.868553] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.868553] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 729.870159] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.870159] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 729.870159] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ebff40d5-c9d8-4560-91fd-2d82ee0fa443, please check neutron logs for more information. [ 729.870159] env[62914]: ERROR nova.compute.manager [ 729.870159] env[62914]: Traceback (most recent call last): [ 729.870159] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 729.870159] env[62914]: listener.cb(fileno) [ 729.870159] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.870159] env[62914]: result = function(*args, **kwargs) [ 729.870159] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 729.870159] env[62914]: return func(*args, **kwargs) [ 729.870159] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.870159] env[62914]: raise e [ 729.870159] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.870159] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 729.870159] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.870159] env[62914]: created_port_ids = self._update_ports_for_instance( [ 729.870159] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.870159] env[62914]: with excutils.save_and_reraise_exception(): [ 729.870159] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.870159] env[62914]: self.force_reraise() [ 729.870159] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.870159] env[62914]: raise self.value [ 729.870159] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.870159] env[62914]: updated_port = self._update_port( [ 729.870159] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.870159] env[62914]: _ensure_no_port_binding_failure(port) [ 729.870159] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.870159] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 729.871087] env[62914]: nova.exception.PortBindingFailed: Binding failed for port ebff40d5-c9d8-4560-91fd-2d82ee0fa443, please check neutron logs for more information. [ 729.871087] env[62914]: Removing descriptor: 19 [ 729.896825] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 729.927190] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 729.927418] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 729.927564] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 729.927734] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 729.927867] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 729.928020] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 729.928452] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 729.928643] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 729.928942] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 729.931020] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 729.931020] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 729.931020] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32971e2-4c56-4f9c-b62c-08b578274c00 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.939785] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-863d0605-5f01-4bbe-abc5-1c01580e628c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.959481] env[62914]: ERROR nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ebff40d5-c9d8-4560-91fd-2d82ee0fa443, please check neutron logs for more information. [ 729.959481] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Traceback (most recent call last): [ 729.959481] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 729.959481] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] yield resources [ 729.959481] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 729.959481] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] self.driver.spawn(context, instance, image_meta, [ 729.959481] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 729.959481] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] self._vmops.spawn(context, instance, image_meta, injected_files, [ 729.959481] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 729.959481] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] vm_ref = self.build_virtual_machine(instance, [ 729.959481] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 729.961918] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] vif_infos = vmwarevif.get_vif_info(self._session, [ 729.961918] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 729.961918] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] for vif in network_info: [ 729.961918] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 729.961918] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] return self._sync_wrapper(fn, *args, **kwargs) [ 729.961918] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 729.961918] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] self.wait() [ 729.961918] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 729.961918] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] self[:] = self._gt.wait() [ 729.961918] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 729.961918] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] return self._exit_event.wait() [ 729.961918] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 729.961918] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] current.throw(*self._exc) [ 729.962428] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.962428] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] result = function(*args, **kwargs) [ 729.962428] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 729.962428] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] return func(*args, **kwargs) [ 729.962428] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.962428] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] raise e [ 729.962428] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.962428] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] nwinfo = self.network_api.allocate_for_instance( [ 729.962428] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.962428] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] created_port_ids = self._update_ports_for_instance( [ 729.962428] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.962428] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] with excutils.save_and_reraise_exception(): [ 729.962428] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.962851] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] self.force_reraise() [ 729.962851] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.962851] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] raise self.value [ 729.962851] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.962851] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] updated_port = self._update_port( [ 729.962851] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.962851] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] _ensure_no_port_binding_failure(port) [ 729.962851] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.962851] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] raise exception.PortBindingFailed(port_id=port['id']) [ 729.962851] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] nova.exception.PortBindingFailed: Binding failed for port ebff40d5-c9d8-4560-91fd-2d82ee0fa443, please check neutron logs for more information. [ 729.962851] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] [ 729.962851] env[62914]: INFO nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Terminating instance [ 729.966452] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "refresh_cache-4c1016c4-2da8-4f4e-9497-4d680fb74535" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.144744] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.187783] env[62914]: DEBUG nova.network.neutron [req-57badbee-32fe-4137-8a42-d4234a0f8163 req-7fa73080-dbfe-402c-914c-f7f13a02c6f7 service nova] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.275520] env[62914]: DEBUG nova.network.neutron [req-57badbee-32fe-4137-8a42-d4234a0f8163 req-7fa73080-dbfe-402c-914c-f7f13a02c6f7 service nova] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.294751] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.414s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.295300] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 730.298047] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.254s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.299307] env[62914]: INFO nova.compute.claims [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 730.778638] env[62914]: DEBUG oslo_concurrency.lockutils [req-57badbee-32fe-4137-8a42-d4234a0f8163 req-7fa73080-dbfe-402c-914c-f7f13a02c6f7 service nova] Releasing lock "refresh_cache-4c1016c4-2da8-4f4e-9497-4d680fb74535" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.779129] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquired lock "refresh_cache-4c1016c4-2da8-4f4e-9497-4d680fb74535" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.779213] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.802990] env[62914]: DEBUG nova.compute.utils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 730.808020] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 730.808020] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 730.862223] env[62914]: DEBUG nova.policy [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd9f93f76e6db40a3924a5429ab2a23f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b6b4c54c48514783a044a75dcdea02b8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 731.312640] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 731.321731] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.369292] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Successfully created port: 908ec2a7-3e08-4b06-8152-3c414c106cbb {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 731.498525] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.706177] env[62914]: DEBUG nova.compute.manager [req-232545a3-0dc2-4554-836b-d54616c6ddd4 req-6e9d0b6d-cea0-455d-b3e8-740774488dc9 service nova] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Received event network-vif-deleted-ebff40d5-c9d8-4560-91fd-2d82ee0fa443 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.739186] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430c6e03-7813-44ca-8ba8-f4a02f6de79c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.747411] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0718068a-91ce-452a-96f5-7481bb8a582a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.784305] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478744cb-065b-47c3-9e4f-6fea215a183e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.791802] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df474de-bf50-4205-bae3-75f935c8d63a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.805859] env[62914]: DEBUG nova.compute.provider_tree [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.003593] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Releasing lock "refresh_cache-4c1016c4-2da8-4f4e-9497-4d680fb74535" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.004158] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 732.004286] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 732.004587] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a025d03a-69b5-415b-846e-31c3a20e08e3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.013911] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70bd6366-afc1-44f5-9cd0-868deb4a188e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.035293] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4c1016c4-2da8-4f4e-9497-4d680fb74535 could not be found. [ 732.035603] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 732.035802] env[62914]: INFO nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Took 0.03 seconds to destroy the instance on the hypervisor. [ 732.036120] env[62914]: DEBUG oslo.service.loopingcall [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.036344] env[62914]: DEBUG nova.compute.manager [-] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 732.036498] env[62914]: DEBUG nova.network.neutron [-] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.066497] env[62914]: DEBUG nova.network.neutron [-] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.310455] env[62914]: DEBUG nova.scheduler.client.report [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.326381] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 732.354188] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 732.354438] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 732.354593] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 732.354757] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 732.354918] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 732.355318] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 732.355517] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 732.355754] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 732.355847] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 732.356190] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 732.356374] env[62914]: DEBUG nova.virt.hardware [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.357258] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780f0cbf-d584-4ffc-be49-7860727c9bb9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.365992] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295449e7-99e9-4d4e-b72f-30ba06fa16b9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.567076] env[62914]: DEBUG nova.network.neutron [-] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.722722] env[62914]: ERROR nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 908ec2a7-3e08-4b06-8152-3c414c106cbb, please check neutron logs for more information. [ 732.722722] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 732.722722] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.722722] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 732.722722] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.722722] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 732.722722] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.722722] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 732.722722] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.722722] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 732.722722] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.722722] env[62914]: ERROR nova.compute.manager raise self.value [ 732.722722] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.722722] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 732.722722] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.722722] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 732.723304] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.723304] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 732.723304] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 908ec2a7-3e08-4b06-8152-3c414c106cbb, please check neutron logs for more information. [ 732.723304] env[62914]: ERROR nova.compute.manager [ 732.723304] env[62914]: Traceback (most recent call last): [ 732.723304] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 732.723304] env[62914]: listener.cb(fileno) [ 732.723304] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.723304] env[62914]: result = function(*args, **kwargs) [ 732.723304] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.723304] env[62914]: return func(*args, **kwargs) [ 732.723304] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.723304] env[62914]: raise e [ 732.723304] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.723304] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 732.723304] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.723304] env[62914]: created_port_ids = self._update_ports_for_instance( [ 732.723304] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.723304] env[62914]: with excutils.save_and_reraise_exception(): [ 732.723304] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.723304] env[62914]: self.force_reraise() [ 732.723304] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.723304] env[62914]: raise self.value [ 732.723304] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.723304] env[62914]: updated_port = self._update_port( [ 732.723304] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.723304] env[62914]: _ensure_no_port_binding_failure(port) [ 732.723304] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.723304] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 732.724228] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 908ec2a7-3e08-4b06-8152-3c414c106cbb, please check neutron logs for more information. [ 732.724228] env[62914]: Removing descriptor: 19 [ 732.724228] env[62914]: ERROR nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 908ec2a7-3e08-4b06-8152-3c414c106cbb, please check neutron logs for more information. [ 732.724228] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Traceback (most recent call last): [ 732.724228] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 732.724228] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] yield resources [ 732.724228] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.724228] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] self.driver.spawn(context, instance, image_meta, [ 732.724228] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 732.724228] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.724228] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.724228] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] vm_ref = self.build_virtual_machine(instance, [ 732.724626] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.724626] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.724626] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.724626] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] for vif in network_info: [ 732.724626] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.724626] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] return self._sync_wrapper(fn, *args, **kwargs) [ 732.724626] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.724626] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] self.wait() [ 732.724626] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.724626] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] self[:] = self._gt.wait() [ 732.724626] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.724626] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] return self._exit_event.wait() [ 732.724626] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 732.725034] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] result = hub.switch() [ 732.725034] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 732.725034] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] return self.greenlet.switch() [ 732.725034] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.725034] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] result = function(*args, **kwargs) [ 732.725034] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.725034] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] return func(*args, **kwargs) [ 732.725034] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.725034] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] raise e [ 732.725034] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.725034] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] nwinfo = self.network_api.allocate_for_instance( [ 732.725034] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.725034] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] created_port_ids = self._update_ports_for_instance( [ 732.725447] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.725447] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] with excutils.save_and_reraise_exception(): [ 732.725447] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.725447] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] self.force_reraise() [ 732.725447] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.725447] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] raise self.value [ 732.725447] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.725447] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] updated_port = self._update_port( [ 732.725447] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.725447] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] _ensure_no_port_binding_failure(port) [ 732.725447] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.725447] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] raise exception.PortBindingFailed(port_id=port['id']) [ 732.726728] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] nova.exception.PortBindingFailed: Binding failed for port 908ec2a7-3e08-4b06-8152-3c414c106cbb, please check neutron logs for more information. [ 732.726728] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] [ 732.726728] env[62914]: INFO nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Terminating instance [ 732.726728] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "refresh_cache-41218b72-0b79-496a-855f-57a972dabd0b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.726941] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquired lock "refresh_cache-41218b72-0b79-496a-855f-57a972dabd0b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.726941] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.820082] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.519s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.820082] env[62914]: DEBUG nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 732.820491] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.798s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.822332] env[62914]: INFO nova.compute.claims [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.046432] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "9fec3e97-4203-4a90-8a43-600d11d7e7a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.046653] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "9fec3e97-4203-4a90-8a43-600d11d7e7a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.070954] env[62914]: INFO nova.compute.manager [-] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Took 1.03 seconds to deallocate network for instance. [ 733.072948] env[62914]: DEBUG nova.compute.claims [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 733.073138] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.247670] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.330022] env[62914]: DEBUG nova.compute.utils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.330022] env[62914]: DEBUG nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 733.330022] env[62914]: DEBUG nova.network.neutron [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 733.337878] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.378704] env[62914]: DEBUG nova.policy [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e170de89bfe49d386e764c112420c92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd997cdf2e1740b29ad51bbe107051aa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 733.707106] env[62914]: DEBUG nova.network.neutron [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Successfully created port: 7c98c7f3-faf2-4254-b99e-3d2381fe8a24 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 733.729884] env[62914]: DEBUG nova.compute.manager [req-933ba34f-9c15-4b18-a212-f91054831598 req-47a53c02-1206-404a-a9a2-b250d6325e10 service nova] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Received event network-changed-908ec2a7-3e08-4b06-8152-3c414c106cbb {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 733.730070] env[62914]: DEBUG nova.compute.manager [req-933ba34f-9c15-4b18-a212-f91054831598 req-47a53c02-1206-404a-a9a2-b250d6325e10 service nova] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Refreshing instance network info cache due to event network-changed-908ec2a7-3e08-4b06-8152-3c414c106cbb. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 733.730246] env[62914]: DEBUG oslo_concurrency.lockutils [req-933ba34f-9c15-4b18-a212-f91054831598 req-47a53c02-1206-404a-a9a2-b250d6325e10 service nova] Acquiring lock "refresh_cache-41218b72-0b79-496a-855f-57a972dabd0b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.832559] env[62914]: DEBUG nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 733.846077] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Releasing lock "refresh_cache-41218b72-0b79-496a-855f-57a972dabd0b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.846461] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 733.846644] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 733.847831] env[62914]: DEBUG oslo_concurrency.lockutils [req-933ba34f-9c15-4b18-a212-f91054831598 req-47a53c02-1206-404a-a9a2-b250d6325e10 service nova] Acquired lock "refresh_cache-41218b72-0b79-496a-855f-57a972dabd0b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.849035] env[62914]: DEBUG nova.network.neutron [req-933ba34f-9c15-4b18-a212-f91054831598 req-47a53c02-1206-404a-a9a2-b250d6325e10 service nova] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Refreshing network info cache for port 908ec2a7-3e08-4b06-8152-3c414c106cbb {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 733.849983] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f19e0ecf-201e-4c3c-a881-37595aeb26e4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.863606] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e80f81-f259-43a1-afbe-4e7b1c581e7e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.888674] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 41218b72-0b79-496a-855f-57a972dabd0b could not be found. [ 733.888782] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 733.888899] env[62914]: INFO nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 733.889143] env[62914]: DEBUG oslo.service.loopingcall [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.889356] env[62914]: DEBUG nova.compute.manager [-] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.889443] env[62914]: DEBUG nova.network.neutron [-] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.910816] env[62914]: DEBUG nova.network.neutron [-] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.140418] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20093a97-38ce-48c7-a031-e6e730a39599 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.147811] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70636bb-a5bc-44b1-891f-d0d8f5c492be {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.177177] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa802ba2-bba8-4629-be60-8d3998967374 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.184609] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac78f9b0-e2d8-4e55-a03d-d52663cefba7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.197694] env[62914]: DEBUG nova.compute.provider_tree [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.372900] env[62914]: DEBUG nova.network.neutron [req-933ba34f-9c15-4b18-a212-f91054831598 req-47a53c02-1206-404a-a9a2-b250d6325e10 service nova] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.412381] env[62914]: DEBUG nova.network.neutron [-] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.420500] env[62914]: DEBUG nova.compute.manager [req-4152189d-520c-4178-862f-ba3a19cef98d req-08a13d26-2269-45a8-b8d7-24f5b96046ba service nova] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Received event network-changed-7c98c7f3-faf2-4254-b99e-3d2381fe8a24 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 734.420686] env[62914]: DEBUG nova.compute.manager [req-4152189d-520c-4178-862f-ba3a19cef98d req-08a13d26-2269-45a8-b8d7-24f5b96046ba service nova] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Refreshing instance network info cache due to event network-changed-7c98c7f3-faf2-4254-b99e-3d2381fe8a24. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 734.421297] env[62914]: DEBUG oslo_concurrency.lockutils [req-4152189d-520c-4178-862f-ba3a19cef98d req-08a13d26-2269-45a8-b8d7-24f5b96046ba service nova] Acquiring lock "refresh_cache-d9057a9e-988f-44ee-9339-ca9a1429378e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.421446] env[62914]: DEBUG oslo_concurrency.lockutils [req-4152189d-520c-4178-862f-ba3a19cef98d req-08a13d26-2269-45a8-b8d7-24f5b96046ba service nova] Acquired lock "refresh_cache-d9057a9e-988f-44ee-9339-ca9a1429378e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.421602] env[62914]: DEBUG nova.network.neutron [req-4152189d-520c-4178-862f-ba3a19cef98d req-08a13d26-2269-45a8-b8d7-24f5b96046ba service nova] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Refreshing network info cache for port 7c98c7f3-faf2-4254-b99e-3d2381fe8a24 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 734.459751] env[62914]: DEBUG nova.network.neutron [req-933ba34f-9c15-4b18-a212-f91054831598 req-47a53c02-1206-404a-a9a2-b250d6325e10 service nova] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.620124] env[62914]: ERROR nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7c98c7f3-faf2-4254-b99e-3d2381fe8a24, please check neutron logs for more information. [ 734.620124] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 734.620124] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.620124] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 734.620124] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.620124] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 734.620124] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.620124] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 734.620124] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.620124] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 734.620124] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.620124] env[62914]: ERROR nova.compute.manager raise self.value [ 734.620124] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.620124] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 734.620124] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.620124] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 734.620649] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.620649] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 734.620649] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7c98c7f3-faf2-4254-b99e-3d2381fe8a24, please check neutron logs for more information. [ 734.620649] env[62914]: ERROR nova.compute.manager [ 734.620649] env[62914]: Traceback (most recent call last): [ 734.620649] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 734.620649] env[62914]: listener.cb(fileno) [ 734.620649] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.620649] env[62914]: result = function(*args, **kwargs) [ 734.620649] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.620649] env[62914]: return func(*args, **kwargs) [ 734.620649] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.620649] env[62914]: raise e [ 734.620649] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.620649] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 734.620649] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.620649] env[62914]: created_port_ids = self._update_ports_for_instance( [ 734.620649] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.620649] env[62914]: with excutils.save_and_reraise_exception(): [ 734.620649] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.620649] env[62914]: self.force_reraise() [ 734.620649] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.620649] env[62914]: raise self.value [ 734.620649] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.620649] env[62914]: updated_port = self._update_port( [ 734.620649] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.620649] env[62914]: _ensure_no_port_binding_failure(port) [ 734.620649] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.620649] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 734.621463] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 7c98c7f3-faf2-4254-b99e-3d2381fe8a24, please check neutron logs for more information. [ 734.621463] env[62914]: Removing descriptor: 15 [ 734.703098] env[62914]: DEBUG nova.scheduler.client.report [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.842150] env[62914]: DEBUG nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 734.869610] env[62914]: DEBUG nova.virt.hardware [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 734.870742] env[62914]: DEBUG nova.virt.hardware [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 734.870742] env[62914]: DEBUG nova.virt.hardware [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 734.870742] env[62914]: DEBUG nova.virt.hardware [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 734.870742] env[62914]: DEBUG nova.virt.hardware [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 734.870742] env[62914]: DEBUG nova.virt.hardware [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 734.871115] env[62914]: DEBUG nova.virt.hardware [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 734.871115] env[62914]: DEBUG nova.virt.hardware [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 734.871115] env[62914]: DEBUG nova.virt.hardware [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 734.871115] env[62914]: DEBUG nova.virt.hardware [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 734.871283] env[62914]: DEBUG nova.virt.hardware [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 734.872131] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53be1917-76e0-4d6e-843f-4c2e4786d4f8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.880961] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61cabc54-9bb6-4d88-a727-fdbe32925955 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.895404] env[62914]: ERROR nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7c98c7f3-faf2-4254-b99e-3d2381fe8a24, please check neutron logs for more information. [ 734.895404] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Traceback (most recent call last): [ 734.895404] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 734.895404] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] yield resources [ 734.895404] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.895404] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] self.driver.spawn(context, instance, image_meta, [ 734.895404] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 734.895404] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.895404] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.895404] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] vm_ref = self.build_virtual_machine(instance, [ 734.895404] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.895766] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.895766] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.895766] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] for vif in network_info: [ 734.895766] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 734.895766] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] return self._sync_wrapper(fn, *args, **kwargs) [ 734.895766] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 734.895766] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] self.wait() [ 734.895766] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 734.895766] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] self[:] = self._gt.wait() [ 734.895766] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.895766] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] return self._exit_event.wait() [ 734.895766] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 734.895766] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] current.throw(*self._exc) [ 734.896292] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.896292] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] result = function(*args, **kwargs) [ 734.896292] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.896292] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] return func(*args, **kwargs) [ 734.896292] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.896292] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] raise e [ 734.896292] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.896292] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] nwinfo = self.network_api.allocate_for_instance( [ 734.896292] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.896292] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] created_port_ids = self._update_ports_for_instance( [ 734.896292] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.896292] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] with excutils.save_and_reraise_exception(): [ 734.896292] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.896669] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] self.force_reraise() [ 734.896669] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.896669] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] raise self.value [ 734.896669] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.896669] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] updated_port = self._update_port( [ 734.896669] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.896669] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] _ensure_no_port_binding_failure(port) [ 734.896669] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.896669] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] raise exception.PortBindingFailed(port_id=port['id']) [ 734.896669] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] nova.exception.PortBindingFailed: Binding failed for port 7c98c7f3-faf2-4254-b99e-3d2381fe8a24, please check neutron logs for more information. [ 734.896669] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] [ 734.896669] env[62914]: INFO nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Terminating instance [ 734.898848] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Acquiring lock "refresh_cache-d9057a9e-988f-44ee-9339-ca9a1429378e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.915390] env[62914]: INFO nova.compute.manager [-] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Took 1.03 seconds to deallocate network for instance. [ 734.917693] env[62914]: DEBUG nova.compute.claims [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 734.917905] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.942393] env[62914]: DEBUG nova.network.neutron [req-4152189d-520c-4178-862f-ba3a19cef98d req-08a13d26-2269-45a8-b8d7-24f5b96046ba service nova] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.962121] env[62914]: DEBUG oslo_concurrency.lockutils [req-933ba34f-9c15-4b18-a212-f91054831598 req-47a53c02-1206-404a-a9a2-b250d6325e10 service nova] Releasing lock "refresh_cache-41218b72-0b79-496a-855f-57a972dabd0b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.962370] env[62914]: DEBUG nova.compute.manager [req-933ba34f-9c15-4b18-a212-f91054831598 req-47a53c02-1206-404a-a9a2-b250d6325e10 service nova] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Received event network-vif-deleted-908ec2a7-3e08-4b06-8152-3c414c106cbb {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.028522] env[62914]: DEBUG nova.network.neutron [req-4152189d-520c-4178-862f-ba3a19cef98d req-08a13d26-2269-45a8-b8d7-24f5b96046ba service nova] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.208209] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.208727] env[62914]: DEBUG nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 735.211285] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.325s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.213080] env[62914]: INFO nova.compute.claims [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.531954] env[62914]: DEBUG oslo_concurrency.lockutils [req-4152189d-520c-4178-862f-ba3a19cef98d req-08a13d26-2269-45a8-b8d7-24f5b96046ba service nova] Releasing lock "refresh_cache-d9057a9e-988f-44ee-9339-ca9a1429378e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.532524] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Acquired lock "refresh_cache-d9057a9e-988f-44ee-9339-ca9a1429378e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.532672] env[62914]: DEBUG nova.network.neutron [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 735.717355] env[62914]: DEBUG nova.compute.utils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.723418] env[62914]: DEBUG nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 735.723418] env[62914]: DEBUG nova.network.neutron [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 735.764659] env[62914]: DEBUG nova.policy [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e170de89bfe49d386e764c112420c92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd997cdf2e1740b29ad51bbe107051aa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 736.040053] env[62914]: DEBUG nova.network.neutron [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Successfully created port: 3345d945-7dfe-4801-b734-0c7a799d785a {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 736.047817] env[62914]: DEBUG nova.network.neutron [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.122114] env[62914]: DEBUG nova.network.neutron [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.226017] env[62914]: DEBUG nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 736.450577] env[62914]: DEBUG nova.compute.manager [req-ec952224-bbc8-4084-9c6a-e8804fb42d6c req-4ba68413-850b-4332-86b2-24ca384f884a service nova] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Received event network-vif-deleted-7c98c7f3-faf2-4254-b99e-3d2381fe8a24 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.523270] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edbff512-e6cc-4f61-8d55-288374d2cff9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.531085] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac3c5db-1c7e-48a2-8dc5-35548ed897aa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.563713] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-950bd6e0-9254-4ed5-ad79-d01c35671ccf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.570620] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e0f1714-f624-457e-9dd7-2da3f9e074a8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.584927] env[62914]: DEBUG nova.compute.provider_tree [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.624243] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Releasing lock "refresh_cache-d9057a9e-988f-44ee-9339-ca9a1429378e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.624771] env[62914]: DEBUG nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 736.625258] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 736.625563] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b96b1118-aeaf-4137-b25b-65cbfdf75d83 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.634806] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d37b92-446b-4bf2-84ad-a69d4e1fd9b2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.656967] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d9057a9e-988f-44ee-9339-ca9a1429378e could not be found. [ 736.657268] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 736.657457] env[62914]: INFO nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 736.657690] env[62914]: DEBUG oslo.service.loopingcall [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 736.658068] env[62914]: DEBUG nova.compute.manager [-] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 736.658161] env[62914]: DEBUG nova.network.neutron [-] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 736.672903] env[62914]: DEBUG nova.network.neutron [-] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.951998] env[62914]: ERROR nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3345d945-7dfe-4801-b734-0c7a799d785a, please check neutron logs for more information. [ 736.951998] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 736.951998] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.951998] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 736.951998] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.951998] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 736.951998] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.951998] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 736.951998] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.951998] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 736.951998] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.951998] env[62914]: ERROR nova.compute.manager raise self.value [ 736.951998] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.951998] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 736.951998] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.951998] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 736.952497] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.952497] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 736.952497] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3345d945-7dfe-4801-b734-0c7a799d785a, please check neutron logs for more information. [ 736.952497] env[62914]: ERROR nova.compute.manager [ 736.952497] env[62914]: Traceback (most recent call last): [ 736.952497] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 736.952497] env[62914]: listener.cb(fileno) [ 736.952497] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.952497] env[62914]: result = function(*args, **kwargs) [ 736.952497] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 736.952497] env[62914]: return func(*args, **kwargs) [ 736.952497] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.952497] env[62914]: raise e [ 736.952497] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.952497] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 736.952497] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.952497] env[62914]: created_port_ids = self._update_ports_for_instance( [ 736.952497] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.952497] env[62914]: with excutils.save_and_reraise_exception(): [ 736.952497] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.952497] env[62914]: self.force_reraise() [ 736.952497] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.952497] env[62914]: raise self.value [ 736.952497] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.952497] env[62914]: updated_port = self._update_port( [ 736.952497] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.952497] env[62914]: _ensure_no_port_binding_failure(port) [ 736.952497] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.952497] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 736.953304] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 3345d945-7dfe-4801-b734-0c7a799d785a, please check neutron logs for more information. [ 736.953304] env[62914]: Removing descriptor: 15 [ 737.088300] env[62914]: DEBUG nova.scheduler.client.report [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.113012] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 737.113229] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 737.174911] env[62914]: DEBUG nova.network.neutron [-] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.235440] env[62914]: DEBUG nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 737.259799] env[62914]: DEBUG nova.virt.hardware [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 737.260059] env[62914]: DEBUG nova.virt.hardware [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 737.260232] env[62914]: DEBUG nova.virt.hardware [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 737.260417] env[62914]: DEBUG nova.virt.hardware [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 737.260559] env[62914]: DEBUG nova.virt.hardware [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 737.260728] env[62914]: DEBUG nova.virt.hardware [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 737.260998] env[62914]: DEBUG nova.virt.hardware [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 737.261125] env[62914]: DEBUG nova.virt.hardware [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 737.261274] env[62914]: DEBUG nova.virt.hardware [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 737.261425] env[62914]: DEBUG nova.virt.hardware [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 737.261589] env[62914]: DEBUG nova.virt.hardware [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.262490] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb6b958-1c8f-45ed-aa13-40c13b6c4683 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.270627] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec1c6d0-8de6-4c53-a8d1-93cc5a274fa5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.285996] env[62914]: ERROR nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3345d945-7dfe-4801-b734-0c7a799d785a, please check neutron logs for more information. [ 737.285996] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Traceback (most recent call last): [ 737.285996] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 737.285996] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] yield resources [ 737.285996] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 737.285996] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] self.driver.spawn(context, instance, image_meta, [ 737.285996] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 737.285996] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.285996] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.285996] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] vm_ref = self.build_virtual_machine(instance, [ 737.285996] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.286473] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.286473] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.286473] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] for vif in network_info: [ 737.286473] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.286473] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] return self._sync_wrapper(fn, *args, **kwargs) [ 737.286473] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.286473] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] self.wait() [ 737.286473] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.286473] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] self[:] = self._gt.wait() [ 737.286473] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.286473] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] return self._exit_event.wait() [ 737.286473] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 737.286473] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] current.throw(*self._exc) [ 737.286896] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.286896] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] result = function(*args, **kwargs) [ 737.286896] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 737.286896] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] return func(*args, **kwargs) [ 737.286896] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.286896] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] raise e [ 737.286896] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.286896] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] nwinfo = self.network_api.allocate_for_instance( [ 737.286896] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.286896] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] created_port_ids = self._update_ports_for_instance( [ 737.286896] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.286896] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] with excutils.save_and_reraise_exception(): [ 737.286896] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.287385] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] self.force_reraise() [ 737.287385] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.287385] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] raise self.value [ 737.287385] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.287385] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] updated_port = self._update_port( [ 737.287385] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.287385] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] _ensure_no_port_binding_failure(port) [ 737.287385] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.287385] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] raise exception.PortBindingFailed(port_id=port['id']) [ 737.287385] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] nova.exception.PortBindingFailed: Binding failed for port 3345d945-7dfe-4801-b734-0c7a799d785a, please check neutron logs for more information. [ 737.287385] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] [ 737.287385] env[62914]: INFO nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Terminating instance [ 737.288448] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Acquiring lock "refresh_cache-220578cb-17b1-4b46-8c4a-f676c04fc5e8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.288616] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Acquired lock "refresh_cache-220578cb-17b1-4b46-8c4a-f676c04fc5e8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.288778] env[62914]: DEBUG nova.network.neutron [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 737.592721] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.593284] env[62914]: DEBUG nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 737.596319] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.569s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.619734] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 737.619908] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Starting heal instance info cache {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 737.677142] env[62914]: INFO nova.compute.manager [-] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Took 1.02 seconds to deallocate network for instance. [ 737.679402] env[62914]: DEBUG nova.compute.claims [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 737.679492] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.807120] env[62914]: DEBUG nova.network.neutron [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.895657] env[62914]: DEBUG nova.network.neutron [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.100950] env[62914]: DEBUG nova.compute.utils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 738.104888] env[62914]: DEBUG nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 738.105078] env[62914]: DEBUG nova.network.neutron [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 738.122817] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Didn't find any instances for network info cache update. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 738.123060] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.123231] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.123378] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.123522] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.123858] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.124604] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.124604] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62914) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 738.124604] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager.update_available_resource {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.167891] env[62914]: DEBUG nova.policy [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91a86a586e81458685325d77e4fcf315', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55c502b78436498cad65050d3b588838', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 738.391552] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9dc0d5-13d4-4556-aa02-6f67b988e8f8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.398319] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Releasing lock "refresh_cache-220578cb-17b1-4b46-8c4a-f676c04fc5e8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.398701] env[62914]: DEBUG nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 738.398886] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 738.399506] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f6a56cd-ee28-43dd-8ff8-e2f088d6371e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.401981] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d01ca3-a7c3-4378-90ab-75cc7d75ffb1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.410856] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3982d341-7c8f-4dd2-b89d-3325cf42fc72 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.445430] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daaf1ca2-1dc0-40d3-b022-b04919acb9e3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.455610] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 220578cb-17b1-4b46-8c4a-f676c04fc5e8 could not be found. [ 738.455932] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 738.460022] env[62914]: INFO nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Took 0.06 seconds to destroy the instance on the hypervisor. [ 738.460022] env[62914]: DEBUG oslo.service.loopingcall [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 738.460022] env[62914]: DEBUG nova.compute.manager [-] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 738.460022] env[62914]: DEBUG nova.network.neutron [-] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 738.460851] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f8735e-5f64-4427-afae-b0739ba292eb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.474473] env[62914]: DEBUG nova.compute.provider_tree [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.478121] env[62914]: DEBUG nova.network.neutron [-] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.485249] env[62914]: DEBUG nova.compute.manager [req-8741286d-4ccf-4190-984b-ef2c76328b78 req-bb4f0687-222b-4c27-8283-df9f14fc7757 service nova] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Received event network-changed-3345d945-7dfe-4801-b734-0c7a799d785a {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.485426] env[62914]: DEBUG nova.compute.manager [req-8741286d-4ccf-4190-984b-ef2c76328b78 req-bb4f0687-222b-4c27-8283-df9f14fc7757 service nova] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Refreshing instance network info cache due to event network-changed-3345d945-7dfe-4801-b734-0c7a799d785a. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 738.485629] env[62914]: DEBUG oslo_concurrency.lockutils [req-8741286d-4ccf-4190-984b-ef2c76328b78 req-bb4f0687-222b-4c27-8283-df9f14fc7757 service nova] Acquiring lock "refresh_cache-220578cb-17b1-4b46-8c4a-f676c04fc5e8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.485764] env[62914]: DEBUG oslo_concurrency.lockutils [req-8741286d-4ccf-4190-984b-ef2c76328b78 req-bb4f0687-222b-4c27-8283-df9f14fc7757 service nova] Acquired lock "refresh_cache-220578cb-17b1-4b46-8c4a-f676c04fc5e8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.485914] env[62914]: DEBUG nova.network.neutron [req-8741286d-4ccf-4190-984b-ef2c76328b78 req-bb4f0687-222b-4c27-8283-df9f14fc7757 service nova] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Refreshing network info cache for port 3345d945-7dfe-4801-b734-0c7a799d785a {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 738.513715] env[62914]: DEBUG nova.network.neutron [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Successfully created port: 385e4c37-f5d7-4035-b356-34fe317498c4 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 738.607732] env[62914]: DEBUG nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 738.627755] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.977617] env[62914]: DEBUG nova.scheduler.client.report [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.981172] env[62914]: DEBUG nova.network.neutron [-] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.003652] env[62914]: DEBUG nova.network.neutron [req-8741286d-4ccf-4190-984b-ef2c76328b78 req-bb4f0687-222b-4c27-8283-df9f14fc7757 service nova] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.101432] env[62914]: DEBUG nova.network.neutron [req-8741286d-4ccf-4190-984b-ef2c76328b78 req-bb4f0687-222b-4c27-8283-df9f14fc7757 service nova] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.444645] env[62914]: ERROR nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 385e4c37-f5d7-4035-b356-34fe317498c4, please check neutron logs for more information. [ 739.444645] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 739.444645] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.444645] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 739.444645] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.444645] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 739.444645] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.444645] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 739.444645] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.444645] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 739.444645] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.444645] env[62914]: ERROR nova.compute.manager raise self.value [ 739.444645] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.444645] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 739.444645] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.444645] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 739.445226] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.445226] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 739.445226] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 385e4c37-f5d7-4035-b356-34fe317498c4, please check neutron logs for more information. [ 739.445226] env[62914]: ERROR nova.compute.manager [ 739.445226] env[62914]: Traceback (most recent call last): [ 739.445226] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 739.445226] env[62914]: listener.cb(fileno) [ 739.445226] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.445226] env[62914]: result = function(*args, **kwargs) [ 739.445226] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 739.445226] env[62914]: return func(*args, **kwargs) [ 739.445226] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.445226] env[62914]: raise e [ 739.445226] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.445226] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 739.445226] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.445226] env[62914]: created_port_ids = self._update_ports_for_instance( [ 739.445226] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.445226] env[62914]: with excutils.save_and_reraise_exception(): [ 739.445226] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.445226] env[62914]: self.force_reraise() [ 739.445226] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.445226] env[62914]: raise self.value [ 739.445226] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.445226] env[62914]: updated_port = self._update_port( [ 739.445226] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.445226] env[62914]: _ensure_no_port_binding_failure(port) [ 739.445226] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.445226] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 739.446217] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 385e4c37-f5d7-4035-b356-34fe317498c4, please check neutron logs for more information. [ 739.446217] env[62914]: Removing descriptor: 15 [ 739.483886] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.887s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.484568] env[62914]: ERROR nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7b798aef-dc26-44e5-b96c-1c00a7e2fe61, please check neutron logs for more information. [ 739.484568] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Traceback (most recent call last): [ 739.484568] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.484568] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] self.driver.spawn(context, instance, image_meta, [ 739.484568] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 739.484568] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.484568] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.484568] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] vm_ref = self.build_virtual_machine(instance, [ 739.484568] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.484568] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.484568] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.484917] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] for vif in network_info: [ 739.484917] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.484917] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] return self._sync_wrapper(fn, *args, **kwargs) [ 739.484917] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.484917] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] self.wait() [ 739.484917] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.484917] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] self[:] = self._gt.wait() [ 739.484917] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.484917] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] return self._exit_event.wait() [ 739.484917] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 739.484917] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] result = hub.switch() [ 739.484917] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 739.484917] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] return self.greenlet.switch() [ 739.485429] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.485429] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] result = function(*args, **kwargs) [ 739.485429] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 739.485429] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] return func(*args, **kwargs) [ 739.485429] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.485429] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] raise e [ 739.485429] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.485429] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] nwinfo = self.network_api.allocate_for_instance( [ 739.485429] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.485429] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] created_port_ids = self._update_ports_for_instance( [ 739.485429] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.485429] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] with excutils.save_and_reraise_exception(): [ 739.485429] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.485782] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] self.force_reraise() [ 739.485782] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.485782] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] raise self.value [ 739.485782] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.485782] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] updated_port = self._update_port( [ 739.485782] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.485782] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] _ensure_no_port_binding_failure(port) [ 739.485782] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.485782] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] raise exception.PortBindingFailed(port_id=port['id']) [ 739.485782] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] nova.exception.PortBindingFailed: Binding failed for port 7b798aef-dc26-44e5-b96c-1c00a7e2fe61, please check neutron logs for more information. [ 739.485782] env[62914]: ERROR nova.compute.manager [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] [ 739.486097] env[62914]: DEBUG nova.compute.utils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Binding failed for port 7b798aef-dc26-44e5-b96c-1c00a7e2fe61, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 739.487016] env[62914]: INFO nova.compute.manager [-] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Took 1.03 seconds to deallocate network for instance. [ 739.487364] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.753s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.491902] env[62914]: DEBUG nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Build of instance 9edfad51-10a5-4e0c-a337-2de909c361cb was re-scheduled: Binding failed for port 7b798aef-dc26-44e5-b96c-1c00a7e2fe61, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 739.492349] env[62914]: DEBUG nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 739.492569] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Acquiring lock "refresh_cache-9edfad51-10a5-4e0c-a337-2de909c361cb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.492711] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Acquired lock "refresh_cache-9edfad51-10a5-4e0c-a337-2de909c361cb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.492862] env[62914]: DEBUG nova.network.neutron [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.502710] env[62914]: DEBUG nova.compute.claims [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 739.502888] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.603854] env[62914]: DEBUG oslo_concurrency.lockutils [req-8741286d-4ccf-4190-984b-ef2c76328b78 req-bb4f0687-222b-4c27-8283-df9f14fc7757 service nova] Releasing lock "refresh_cache-220578cb-17b1-4b46-8c4a-f676c04fc5e8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.604132] env[62914]: DEBUG nova.compute.manager [req-8741286d-4ccf-4190-984b-ef2c76328b78 req-bb4f0687-222b-4c27-8283-df9f14fc7757 service nova] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Received event network-vif-deleted-3345d945-7dfe-4801-b734-0c7a799d785a {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 739.617539] env[62914]: DEBUG nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 739.641485] env[62914]: DEBUG nova.virt.hardware [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 739.641740] env[62914]: DEBUG nova.virt.hardware [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 739.641892] env[62914]: DEBUG nova.virt.hardware [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 739.642090] env[62914]: DEBUG nova.virt.hardware [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 739.642237] env[62914]: DEBUG nova.virt.hardware [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 739.642380] env[62914]: DEBUG nova.virt.hardware [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 739.642577] env[62914]: DEBUG nova.virt.hardware [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 739.642730] env[62914]: DEBUG nova.virt.hardware [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 739.642891] env[62914]: DEBUG nova.virt.hardware [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 739.643065] env[62914]: DEBUG nova.virt.hardware [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 739.643251] env[62914]: DEBUG nova.virt.hardware [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 739.644117] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50b466c-400a-4c42-910f-fcf9fd31f455 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.652905] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974acb08-dbeb-432b-99ab-d6c5bfc06fe7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.668237] env[62914]: ERROR nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 385e4c37-f5d7-4035-b356-34fe317498c4, please check neutron logs for more information. [ 739.668237] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Traceback (most recent call last): [ 739.668237] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 739.668237] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] yield resources [ 739.668237] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.668237] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] self.driver.spawn(context, instance, image_meta, [ 739.668237] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 739.668237] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.668237] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.668237] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] vm_ref = self.build_virtual_machine(instance, [ 739.668237] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.668620] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.668620] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.668620] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] for vif in network_info: [ 739.668620] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.668620] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] return self._sync_wrapper(fn, *args, **kwargs) [ 739.668620] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.668620] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] self.wait() [ 739.668620] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.668620] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] self[:] = self._gt.wait() [ 739.668620] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.668620] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] return self._exit_event.wait() [ 739.668620] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 739.668620] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] current.throw(*self._exc) [ 739.669199] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.669199] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] result = function(*args, **kwargs) [ 739.669199] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 739.669199] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] return func(*args, **kwargs) [ 739.669199] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.669199] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] raise e [ 739.669199] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.669199] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] nwinfo = self.network_api.allocate_for_instance( [ 739.669199] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.669199] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] created_port_ids = self._update_ports_for_instance( [ 739.669199] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.669199] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] with excutils.save_and_reraise_exception(): [ 739.669199] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.669790] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] self.force_reraise() [ 739.669790] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.669790] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] raise self.value [ 739.669790] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.669790] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] updated_port = self._update_port( [ 739.669790] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.669790] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] _ensure_no_port_binding_failure(port) [ 739.669790] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.669790] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] raise exception.PortBindingFailed(port_id=port['id']) [ 739.669790] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] nova.exception.PortBindingFailed: Binding failed for port 385e4c37-f5d7-4035-b356-34fe317498c4, please check neutron logs for more information. [ 739.669790] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] [ 739.669790] env[62914]: INFO nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Terminating instance [ 739.670451] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Acquiring lock "refresh_cache-d73f5ea8-529a-44e3-8239-5aaa4534d1cc" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.670607] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Acquired lock "refresh_cache-d73f5ea8-529a-44e3-8239-5aaa4534d1cc" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.670767] env[62914]: DEBUG nova.network.neutron [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.010065] env[62914]: DEBUG nova.network.neutron [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.107164] env[62914]: DEBUG nova.network.neutron [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.189816] env[62914]: DEBUG nova.network.neutron [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.236379] env[62914]: DEBUG nova.network.neutron [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.283427] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0155cc-7770-4d09-9d2b-3568d95c356e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.291072] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06faa943-d814-4b62-befd-b9eb546eb86a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.321011] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a83a3bf-5e17-4de5-962d-f0e1ef3edfe8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.327848] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474e3e59-dddf-4453-a781-27ff34ef4cc1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.340475] env[62914]: DEBUG nova.compute.provider_tree [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.506594] env[62914]: DEBUG nova.compute.manager [req-bfd31c5c-a72d-4c21-9832-ae5783be2a73 req-8583dff7-a545-4114-bceb-5f7812227d90 service nova] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Received event network-changed-385e4c37-f5d7-4035-b356-34fe317498c4 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.506695] env[62914]: DEBUG nova.compute.manager [req-bfd31c5c-a72d-4c21-9832-ae5783be2a73 req-8583dff7-a545-4114-bceb-5f7812227d90 service nova] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Refreshing instance network info cache due to event network-changed-385e4c37-f5d7-4035-b356-34fe317498c4. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 740.506872] env[62914]: DEBUG oslo_concurrency.lockutils [req-bfd31c5c-a72d-4c21-9832-ae5783be2a73 req-8583dff7-a545-4114-bceb-5f7812227d90 service nova] Acquiring lock "refresh_cache-d73f5ea8-529a-44e3-8239-5aaa4534d1cc" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.608328] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Releasing lock "refresh_cache-9edfad51-10a5-4e0c-a337-2de909c361cb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.608566] env[62914]: DEBUG nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 740.608748] env[62914]: DEBUG nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.608905] env[62914]: DEBUG nova.network.neutron [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.623629] env[62914]: DEBUG nova.network.neutron [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.739667] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Releasing lock "refresh_cache-d73f5ea8-529a-44e3-8239-5aaa4534d1cc" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.740124] env[62914]: DEBUG nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 740.740357] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 740.740663] env[62914]: DEBUG oslo_concurrency.lockutils [req-bfd31c5c-a72d-4c21-9832-ae5783be2a73 req-8583dff7-a545-4114-bceb-5f7812227d90 service nova] Acquired lock "refresh_cache-d73f5ea8-529a-44e3-8239-5aaa4534d1cc" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.740829] env[62914]: DEBUG nova.network.neutron [req-bfd31c5c-a72d-4c21-9832-ae5783be2a73 req-8583dff7-a545-4114-bceb-5f7812227d90 service nova] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Refreshing network info cache for port 385e4c37-f5d7-4035-b356-34fe317498c4 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 740.741865] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4a6d69a5-20c1-4fde-ac80-24ad862e8283 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.751676] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0caa42f-1ab9-4e46-b172-35e8b7f315d5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.774585] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d73f5ea8-529a-44e3-8239-5aaa4534d1cc could not be found. [ 740.774803] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 740.774977] env[62914]: INFO nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Took 0.03 seconds to destroy the instance on the hypervisor. [ 740.775231] env[62914]: DEBUG oslo.service.loopingcall [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.775443] env[62914]: DEBUG nova.compute.manager [-] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.775538] env[62914]: DEBUG nova.network.neutron [-] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.791190] env[62914]: DEBUG nova.network.neutron [-] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.843549] env[62914]: DEBUG nova.scheduler.client.report [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.126838] env[62914]: DEBUG nova.network.neutron [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.259456] env[62914]: DEBUG nova.network.neutron [req-bfd31c5c-a72d-4c21-9832-ae5783be2a73 req-8583dff7-a545-4114-bceb-5f7812227d90 service nova] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.292803] env[62914]: DEBUG nova.network.neutron [-] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.302709] env[62914]: DEBUG nova.network.neutron [req-bfd31c5c-a72d-4c21-9832-ae5783be2a73 req-8583dff7-a545-4114-bceb-5f7812227d90 service nova] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.350034] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.863s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.350887] env[62914]: ERROR nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1195773f-b43f-4e3a-944c-f6116a5b8ac7, please check neutron logs for more information. [ 741.350887] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Traceback (most recent call last): [ 741.350887] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 741.350887] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] self.driver.spawn(context, instance, image_meta, [ 741.350887] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 741.350887] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 741.350887] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 741.350887] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] vm_ref = self.build_virtual_machine(instance, [ 741.350887] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 741.350887] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] vif_infos = vmwarevif.get_vif_info(self._session, [ 741.350887] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 741.351293] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] for vif in network_info: [ 741.351293] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 741.351293] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] return self._sync_wrapper(fn, *args, **kwargs) [ 741.351293] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 741.351293] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] self.wait() [ 741.351293] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 741.351293] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] self[:] = self._gt.wait() [ 741.351293] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 741.351293] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] return self._exit_event.wait() [ 741.351293] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 741.351293] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] result = hub.switch() [ 741.351293] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 741.351293] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] return self.greenlet.switch() [ 741.351768] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.351768] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] result = function(*args, **kwargs) [ 741.351768] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 741.351768] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] return func(*args, **kwargs) [ 741.351768] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.351768] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] raise e [ 741.351768] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.351768] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] nwinfo = self.network_api.allocate_for_instance( [ 741.351768] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 741.351768] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] created_port_ids = self._update_ports_for_instance( [ 741.351768] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 741.351768] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] with excutils.save_and_reraise_exception(): [ 741.351768] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.352203] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] self.force_reraise() [ 741.352203] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.352203] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] raise self.value [ 741.352203] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 741.352203] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] updated_port = self._update_port( [ 741.352203] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.352203] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] _ensure_no_port_binding_failure(port) [ 741.352203] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.352203] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] raise exception.PortBindingFailed(port_id=port['id']) [ 741.352203] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] nova.exception.PortBindingFailed: Binding failed for port 1195773f-b43f-4e3a-944c-f6116a5b8ac7, please check neutron logs for more information. [ 741.352203] env[62914]: ERROR nova.compute.manager [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] [ 741.352586] env[62914]: DEBUG nova.compute.utils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Binding failed for port 1195773f-b43f-4e3a-944c-f6116a5b8ac7, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 741.352586] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.540s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.354217] env[62914]: INFO nova.compute.claims [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 741.357918] env[62914]: DEBUG nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Build of instance e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6 was re-scheduled: Binding failed for port 1195773f-b43f-4e3a-944c-f6116a5b8ac7, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 741.358356] env[62914]: DEBUG nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 741.358579] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Acquiring lock "refresh_cache-e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.358719] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Acquired lock "refresh_cache-e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.358872] env[62914]: DEBUG nova.network.neutron [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 741.629396] env[62914]: INFO nova.compute.manager [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: 9edfad51-10a5-4e0c-a337-2de909c361cb] Took 1.02 seconds to deallocate network for instance. [ 741.795295] env[62914]: INFO nova.compute.manager [-] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Took 1.02 seconds to deallocate network for instance. [ 741.798063] env[62914]: DEBUG nova.compute.claims [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 741.798308] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.804827] env[62914]: DEBUG oslo_concurrency.lockutils [req-bfd31c5c-a72d-4c21-9832-ae5783be2a73 req-8583dff7-a545-4114-bceb-5f7812227d90 service nova] Releasing lock "refresh_cache-d73f5ea8-529a-44e3-8239-5aaa4534d1cc" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.805108] env[62914]: DEBUG nova.compute.manager [req-bfd31c5c-a72d-4c21-9832-ae5783be2a73 req-8583dff7-a545-4114-bceb-5f7812227d90 service nova] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Received event network-vif-deleted-385e4c37-f5d7-4035-b356-34fe317498c4 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 741.878870] env[62914]: DEBUG nova.network.neutron [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.957819] env[62914]: DEBUG nova.network.neutron [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.460377] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Releasing lock "refresh_cache-e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.460641] env[62914]: DEBUG nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 742.462023] env[62914]: DEBUG nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.462023] env[62914]: DEBUG nova.network.neutron [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 742.477841] env[62914]: DEBUG nova.network.neutron [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.629376] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099c022c-ceb9-46fe-8a11-8420cd1e6163 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.636877] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3618e1-4119-49af-8942-70c1eb809ffe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.669220] env[62914]: INFO nova.scheduler.client.report [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Deleted allocations for instance 9edfad51-10a5-4e0c-a337-2de909c361cb [ 742.674779] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab7a88ae-ab93-48bb-a8e0-6e3cf9c094a7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.682639] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c94a02ff-10e9-41a2-b3ec-d562deb12351 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.695812] env[62914]: DEBUG nova.compute.provider_tree [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.980508] env[62914]: DEBUG nova.network.neutron [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.179553] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a1e9844c-ec08-4a4c-8519-4d53424734ad tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Lock "9edfad51-10a5-4e0c-a337-2de909c361cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.798s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.198641] env[62914]: DEBUG nova.scheduler.client.report [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.483374] env[62914]: INFO nova.compute.manager [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] [instance: e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6] Took 1.02 seconds to deallocate network for instance. [ 743.682043] env[62914]: DEBUG nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 743.703823] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.351s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.704380] env[62914]: DEBUG nova.compute.manager [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 743.707239] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.054s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.206268] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.212032] env[62914]: DEBUG nova.compute.utils [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 744.212948] env[62914]: DEBUG nova.compute.manager [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Not allocating networking since 'none' was specified. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 744.477699] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb383b00-7856-47a4-94b1-c8a04b421b30 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.485110] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fcb55f-540e-49ed-a6bc-3099802ff54a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.517851] env[62914]: INFO nova.scheduler.client.report [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Deleted allocations for instance e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6 [ 744.523367] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-416092b5-4c0b-4516-85f8-50c5128df262 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.531619] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c94a28b-8572-4ea0-8c4d-c1adf960ab4e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.544877] env[62914]: DEBUG nova.compute.provider_tree [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.720793] env[62914]: DEBUG nova.compute.manager [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 745.028753] env[62914]: DEBUG oslo_concurrency.lockutils [None req-99590777-94da-4cb0-947b-c7c5a3b74991 tempest-ServerRescueNegativeTestJSON-794624105 tempest-ServerRescueNegativeTestJSON-794624105-project-member] Lock "e1f45ab1-53fe-417d-b9f1-8ca1e7ce1fa6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.643s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.049979] env[62914]: DEBUG nova.scheduler.client.report [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.531943] env[62914]: DEBUG nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 745.554764] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.847s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.556035] env[62914]: ERROR nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1ac3a328-5cd6-4343-8358-b1ea3e945ea2, please check neutron logs for more information. [ 745.556035] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Traceback (most recent call last): [ 745.556035] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 745.556035] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] self.driver.spawn(context, instance, image_meta, [ 745.556035] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 745.556035] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 745.556035] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 745.556035] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] vm_ref = self.build_virtual_machine(instance, [ 745.556035] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 745.556035] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] vif_infos = vmwarevif.get_vif_info(self._session, [ 745.556035] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 745.556436] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] for vif in network_info: [ 745.556436] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 745.556436] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] return self._sync_wrapper(fn, *args, **kwargs) [ 745.556436] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 745.556436] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] self.wait() [ 745.556436] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 745.556436] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] self[:] = self._gt.wait() [ 745.556436] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 745.556436] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] return self._exit_event.wait() [ 745.556436] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 745.556436] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] current.throw(*self._exc) [ 745.556436] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.556436] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] result = function(*args, **kwargs) [ 745.556813] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 745.556813] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] return func(*args, **kwargs) [ 745.556813] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.556813] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] raise e [ 745.556813] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.556813] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] nwinfo = self.network_api.allocate_for_instance( [ 745.556813] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.556813] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] created_port_ids = self._update_ports_for_instance( [ 745.556813] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.556813] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] with excutils.save_and_reraise_exception(): [ 745.556813] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.556813] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] self.force_reraise() [ 745.556813] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.557302] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] raise self.value [ 745.557302] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.557302] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] updated_port = self._update_port( [ 745.557302] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.557302] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] _ensure_no_port_binding_failure(port) [ 745.557302] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.557302] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] raise exception.PortBindingFailed(port_id=port['id']) [ 745.557302] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] nova.exception.PortBindingFailed: Binding failed for port 1ac3a328-5cd6-4343-8358-b1ea3e945ea2, please check neutron logs for more information. [ 745.557302] env[62914]: ERROR nova.compute.manager [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] [ 745.557302] env[62914]: DEBUG nova.compute.utils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Binding failed for port 1ac3a328-5cd6-4343-8358-b1ea3e945ea2, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 745.557673] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.413s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.559193] env[62914]: INFO nova.compute.claims [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.561799] env[62914]: DEBUG nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Build of instance 8399d066-43bc-4766-97c8-89d74f3283f0 was re-scheduled: Binding failed for port 1ac3a328-5cd6-4343-8358-b1ea3e945ea2, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 745.562249] env[62914]: DEBUG nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 745.562477] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "refresh_cache-8399d066-43bc-4766-97c8-89d74f3283f0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.562622] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquired lock "refresh_cache-8399d066-43bc-4766-97c8-89d74f3283f0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.562775] env[62914]: DEBUG nova.network.neutron [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.731025] env[62914]: DEBUG nova.compute.manager [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 745.757361] env[62914]: DEBUG nova.virt.hardware [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 745.757625] env[62914]: DEBUG nova.virt.hardware [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 745.757789] env[62914]: DEBUG nova.virt.hardware [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 745.757973] env[62914]: DEBUG nova.virt.hardware [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 745.758129] env[62914]: DEBUG nova.virt.hardware [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 745.758295] env[62914]: DEBUG nova.virt.hardware [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 745.758514] env[62914]: DEBUG nova.virt.hardware [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 745.758668] env[62914]: DEBUG nova.virt.hardware [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 745.758999] env[62914]: DEBUG nova.virt.hardware [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 745.758999] env[62914]: DEBUG nova.virt.hardware [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 745.759155] env[62914]: DEBUG nova.virt.hardware [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.760034] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b46d7d4-07f0-4e76-937f-3e99fa12c82c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.767848] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330416c7-2c54-41c9-ae19-ab6de5ef776c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.781544] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Instance VIF info [] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 745.787125] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Creating folder: Project (2efc3ebbb2ce422bb24ab3810b944f23). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 745.787526] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f02666a3-947a-456c-b6ba-7ed0a0b2637c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.797939] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Created folder: Project (2efc3ebbb2ce422bb24ab3810b944f23) in parent group-v288131. [ 745.798236] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Creating folder: Instances. Parent ref: group-v288156. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 745.798421] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1479e3d2-f44a-407e-925e-05c03c0bca4a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.808316] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Created folder: Instances in parent group-v288156. [ 745.808645] env[62914]: DEBUG oslo.service.loopingcall [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.808717] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 745.808914] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-79460229-8877-49eb-9e2c-f78cd13b45b5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.825362] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 745.825362] env[62914]: value = "task-1352414" [ 745.825362] env[62914]: _type = "Task" [ 745.825362] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.835852] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352414, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.055646] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.090824] env[62914]: DEBUG nova.network.neutron [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.238279] env[62914]: DEBUG nova.network.neutron [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.335441] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352414, 'name': CreateVM_Task, 'duration_secs': 0.285558} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.335636] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 746.336073] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.336236] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.336546] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 746.336790] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76c3e38f-e652-4d63-aa56-75d4ba10466a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.341170] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 746.341170] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52516620-0324-3c87-8214-ec65d9f9a076" [ 746.341170] env[62914]: _type = "Task" [ 746.341170] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.348970] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52516620-0324-3c87-8214-ec65d9f9a076, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.740280] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Releasing lock "refresh_cache-8399d066-43bc-4766-97c8-89d74f3283f0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.740641] env[62914]: DEBUG nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 746.741016] env[62914]: DEBUG nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 746.741016] env[62914]: DEBUG nova.network.neutron [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 746.759109] env[62914]: DEBUG nova.network.neutron [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.839983] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27aa4783-5a4c-467f-8702-000f2f2f7335 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.851966] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52516620-0324-3c87-8214-ec65d9f9a076, 'name': SearchDatastore_Task, 'duration_secs': 0.009478} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.854325] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.854436] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 746.856143] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.856143] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.856143] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 746.856143] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-271ae724-5d45-4c22-acd9-9575342ec566 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.857876] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454bb31f-7874-4789-b90b-afd3c08d91fe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.888957] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ab537f-412a-499d-986c-c56450336c37 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.891664] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 746.891719] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 746.892432] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6c10d48-9cce-4ad7-a474-ab8d28d64f9f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.900229] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40086b6-ebb5-43b1-a850-87dffbbcccdd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.904079] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 746.904079] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]528dfb88-f9a2-4905-16ff-c551a3942136" [ 746.904079] env[62914]: _type = "Task" [ 746.904079] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.915199] env[62914]: DEBUG nova.compute.provider_tree [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.921300] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]528dfb88-f9a2-4905-16ff-c551a3942136, 'name': SearchDatastore_Task, 'duration_secs': 0.008779} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.922039] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a8d9f55-f859-40c6-86b1-6b83fe70b863 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.928366] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 746.928366] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]522a4118-254e-a051-fdb0-557156832bfc" [ 746.928366] env[62914]: _type = "Task" [ 746.928366] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.936209] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]522a4118-254e-a051-fdb0-557156832bfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.261513] env[62914]: DEBUG nova.network.neutron [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.418064] env[62914]: DEBUG nova.scheduler.client.report [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 747.441081] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]522a4118-254e-a051-fdb0-557156832bfc, 'name': SearchDatastore_Task, 'duration_secs': 0.011401} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.441363] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.441617] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 44339b45-5183-4882-aaec-4070adc3c3dd/44339b45-5183-4882-aaec-4070adc3c3dd.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 747.441855] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2376f31a-13cf-45d8-891d-576af7b2a813 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.449226] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 747.449226] env[62914]: value = "task-1352415" [ 747.449226] env[62914]: _type = "Task" [ 747.449226] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.457332] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352415, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.764397] env[62914]: INFO nova.compute.manager [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 8399d066-43bc-4766-97c8-89d74f3283f0] Took 1.02 seconds to deallocate network for instance. [ 747.922627] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.923182] env[62914]: DEBUG nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 747.925687] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.853s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.961810] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352415, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486924} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.962126] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 44339b45-5183-4882-aaec-4070adc3c3dd/44339b45-5183-4882-aaec-4070adc3c3dd.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 747.962348] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 747.962594] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8ee3f599-901f-471f-bcbd-c1ed320f9f7f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.969697] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 747.969697] env[62914]: value = "task-1352416" [ 747.969697] env[62914]: _type = "Task" [ 747.969697] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.976818] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352416, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.430794] env[62914]: DEBUG nova.compute.utils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.435756] env[62914]: DEBUG nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 748.435960] env[62914]: DEBUG nova.network.neutron [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 748.483234] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352416, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067804} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.483545] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 748.484443] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f092be-1915-41c8-ae1c-a84545b4db33 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.503795] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 44339b45-5183-4882-aaec-4070adc3c3dd/44339b45-5183-4882-aaec-4070adc3c3dd.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 748.507711] env[62914]: DEBUG nova.policy [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '795b94761b144a1e99ed79838d9687ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9278818cebd842669b802a4e43e91774', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 748.509070] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fde594e-cd76-4290-bf43-8264f6c34e38 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.530510] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 748.530510] env[62914]: value = "task-1352417" [ 748.530510] env[62914]: _type = "Task" [ 748.530510] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.539666] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352417, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.792243] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784dbeb6-e41c-4776-beda-789f69b2f6b8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.797512] env[62914]: INFO nova.scheduler.client.report [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Deleted allocations for instance 8399d066-43bc-4766-97c8-89d74f3283f0 [ 748.812823] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-905975e5-8371-4196-8006-7b1f3b5f6e9c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.853144] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42112258-0597-43e9-aad3-dd2586e73839 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.862027] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b8423e-f126-4dd9-8120-0a73d966d488 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.876789] env[62914]: DEBUG nova.compute.provider_tree [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.879869] env[62914]: DEBUG nova.network.neutron [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Successfully created port: f517bef4-915c-434b-95aa-baf973fd5446 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.940199] env[62914]: DEBUG nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 749.041994] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352417, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.309123] env[62914]: DEBUG oslo_concurrency.lockutils [None req-248d0a4f-faee-480f-9d0e-29351037307c tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "8399d066-43bc-4766-97c8-89d74f3283f0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.598s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.383209] env[62914]: DEBUG nova.scheduler.client.report [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.547358] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352417, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.811675] env[62914]: DEBUG nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 749.887626] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.962s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.888268] env[62914]: ERROR nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ebff40d5-c9d8-4560-91fd-2d82ee0fa443, please check neutron logs for more information. [ 749.888268] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Traceback (most recent call last): [ 749.888268] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.888268] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] self.driver.spawn(context, instance, image_meta, [ 749.888268] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 749.888268] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.888268] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.888268] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] vm_ref = self.build_virtual_machine(instance, [ 749.888268] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.888268] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.888268] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.888635] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] for vif in network_info: [ 749.888635] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.888635] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] return self._sync_wrapper(fn, *args, **kwargs) [ 749.888635] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.888635] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] self.wait() [ 749.888635] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.888635] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] self[:] = self._gt.wait() [ 749.888635] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.888635] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] return self._exit_event.wait() [ 749.888635] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 749.888635] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] current.throw(*self._exc) [ 749.888635] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.888635] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] result = function(*args, **kwargs) [ 749.889064] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.889064] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] return func(*args, **kwargs) [ 749.889064] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.889064] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] raise e [ 749.889064] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.889064] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] nwinfo = self.network_api.allocate_for_instance( [ 749.889064] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.889064] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] created_port_ids = self._update_ports_for_instance( [ 749.889064] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.889064] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] with excutils.save_and_reraise_exception(): [ 749.889064] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.889064] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] self.force_reraise() [ 749.889064] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.889424] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] raise self.value [ 749.889424] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.889424] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] updated_port = self._update_port( [ 749.889424] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.889424] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] _ensure_no_port_binding_failure(port) [ 749.889424] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.889424] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] raise exception.PortBindingFailed(port_id=port['id']) [ 749.889424] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] nova.exception.PortBindingFailed: Binding failed for port ebff40d5-c9d8-4560-91fd-2d82ee0fa443, please check neutron logs for more information. [ 749.889424] env[62914]: ERROR nova.compute.manager [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] [ 749.889424] env[62914]: DEBUG nova.compute.utils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Binding failed for port ebff40d5-c9d8-4560-91fd-2d82ee0fa443, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 749.891685] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.974s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.897783] env[62914]: DEBUG nova.compute.manager [req-03b4b040-0062-41bc-8b40-b269d0c9820b req-810703f8-2461-40e2-9f40-039847c3a180 service nova] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Received event network-changed-f517bef4-915c-434b-95aa-baf973fd5446 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.898043] env[62914]: DEBUG nova.compute.manager [req-03b4b040-0062-41bc-8b40-b269d0c9820b req-810703f8-2461-40e2-9f40-039847c3a180 service nova] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Refreshing instance network info cache due to event network-changed-f517bef4-915c-434b-95aa-baf973fd5446. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 749.898357] env[62914]: DEBUG oslo_concurrency.lockutils [req-03b4b040-0062-41bc-8b40-b269d0c9820b req-810703f8-2461-40e2-9f40-039847c3a180 service nova] Acquiring lock "refresh_cache-98c19874-7c71-4e20-aa29-2ddc53df7e81" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.898583] env[62914]: DEBUG oslo_concurrency.lockutils [req-03b4b040-0062-41bc-8b40-b269d0c9820b req-810703f8-2461-40e2-9f40-039847c3a180 service nova] Acquired lock "refresh_cache-98c19874-7c71-4e20-aa29-2ddc53df7e81" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.898803] env[62914]: DEBUG nova.network.neutron [req-03b4b040-0062-41bc-8b40-b269d0c9820b req-810703f8-2461-40e2-9f40-039847c3a180 service nova] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Refreshing network info cache for port f517bef4-915c-434b-95aa-baf973fd5446 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 749.900326] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Build of instance 4c1016c4-2da8-4f4e-9497-4d680fb74535 was re-scheduled: Binding failed for port ebff40d5-c9d8-4560-91fd-2d82ee0fa443, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 749.902843] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 749.903090] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "refresh_cache-4c1016c4-2da8-4f4e-9497-4d680fb74535" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.903230] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquired lock "refresh_cache-4c1016c4-2da8-4f4e-9497-4d680fb74535" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.903388] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 749.950613] env[62914]: DEBUG nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 749.982608] env[62914]: DEBUG nova.virt.hardware [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 749.982874] env[62914]: DEBUG nova.virt.hardware [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 749.983035] env[62914]: DEBUG nova.virt.hardware [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.983221] env[62914]: DEBUG nova.virt.hardware [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 749.983361] env[62914]: DEBUG nova.virt.hardware [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.983757] env[62914]: DEBUG nova.virt.hardware [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 749.983999] env[62914]: DEBUG nova.virt.hardware [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 749.984174] env[62914]: DEBUG nova.virt.hardware [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 749.984341] env[62914]: DEBUG nova.virt.hardware [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 749.984501] env[62914]: DEBUG nova.virt.hardware [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 749.984667] env[62914]: DEBUG nova.virt.hardware [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.985576] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa9174e-3801-4542-bae2-95281a5a1778 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.994071] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286b00a2-f505-42cf-b747-31cbef427f3c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.042211] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352417, 'name': ReconfigVM_Task, 'duration_secs': 1.155883} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.042498] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 44339b45-5183-4882-aaec-4070adc3c3dd/44339b45-5183-4882-aaec-4070adc3c3dd.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 750.043127] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ed38bd50-6ccb-49ca-8d88-9f2d8fda9b1b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.049501] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 750.049501] env[62914]: value = "task-1352418" [ 750.049501] env[62914]: _type = "Task" [ 750.049501] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.058012] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352418, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.156665] env[62914]: ERROR nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f517bef4-915c-434b-95aa-baf973fd5446, please check neutron logs for more information. [ 750.156665] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 750.156665] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.156665] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 750.156665] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.156665] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 750.156665] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.156665] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 750.156665] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.156665] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 750.156665] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.156665] env[62914]: ERROR nova.compute.manager raise self.value [ 750.156665] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.156665] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 750.156665] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.156665] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 750.157264] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.157264] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 750.157264] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f517bef4-915c-434b-95aa-baf973fd5446, please check neutron logs for more information. [ 750.157264] env[62914]: ERROR nova.compute.manager [ 750.157264] env[62914]: Traceback (most recent call last): [ 750.157264] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 750.157264] env[62914]: listener.cb(fileno) [ 750.157264] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.157264] env[62914]: result = function(*args, **kwargs) [ 750.157264] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 750.157264] env[62914]: return func(*args, **kwargs) [ 750.157264] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.157264] env[62914]: raise e [ 750.157264] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.157264] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 750.157264] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.157264] env[62914]: created_port_ids = self._update_ports_for_instance( [ 750.157264] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.157264] env[62914]: with excutils.save_and_reraise_exception(): [ 750.157264] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.157264] env[62914]: self.force_reraise() [ 750.157264] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.157264] env[62914]: raise self.value [ 750.157264] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.157264] env[62914]: updated_port = self._update_port( [ 750.157264] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.157264] env[62914]: _ensure_no_port_binding_failure(port) [ 750.157264] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.157264] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 750.158248] env[62914]: nova.exception.PortBindingFailed: Binding failed for port f517bef4-915c-434b-95aa-baf973fd5446, please check neutron logs for more information. [ 750.158248] env[62914]: Removing descriptor: 15 [ 750.158248] env[62914]: ERROR nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f517bef4-915c-434b-95aa-baf973fd5446, please check neutron logs for more information. [ 750.158248] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Traceback (most recent call last): [ 750.158248] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 750.158248] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] yield resources [ 750.158248] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 750.158248] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] self.driver.spawn(context, instance, image_meta, [ 750.158248] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 750.158248] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.158248] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.158248] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] vm_ref = self.build_virtual_machine(instance, [ 750.158713] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.158713] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.158713] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.158713] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] for vif in network_info: [ 750.158713] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 750.158713] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] return self._sync_wrapper(fn, *args, **kwargs) [ 750.158713] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 750.158713] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] self.wait() [ 750.158713] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 750.158713] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] self[:] = self._gt.wait() [ 750.158713] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.158713] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] return self._exit_event.wait() [ 750.158713] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 750.159205] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] result = hub.switch() [ 750.159205] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 750.159205] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] return self.greenlet.switch() [ 750.159205] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.159205] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] result = function(*args, **kwargs) [ 750.159205] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 750.159205] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] return func(*args, **kwargs) [ 750.159205] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.159205] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] raise e [ 750.159205] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.159205] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] nwinfo = self.network_api.allocate_for_instance( [ 750.159205] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.159205] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] created_port_ids = self._update_ports_for_instance( [ 750.159638] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.159638] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] with excutils.save_and_reraise_exception(): [ 750.159638] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.159638] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] self.force_reraise() [ 750.159638] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.159638] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] raise self.value [ 750.159638] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.159638] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] updated_port = self._update_port( [ 750.159638] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.159638] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] _ensure_no_port_binding_failure(port) [ 750.159638] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.159638] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] raise exception.PortBindingFailed(port_id=port['id']) [ 750.160057] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] nova.exception.PortBindingFailed: Binding failed for port f517bef4-915c-434b-95aa-baf973fd5446, please check neutron logs for more information. [ 750.160057] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] [ 750.160057] env[62914]: INFO nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Terminating instance [ 750.160057] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "refresh_cache-98c19874-7c71-4e20-aa29-2ddc53df7e81" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.337515] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.424175] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.426660] env[62914]: DEBUG nova.network.neutron [req-03b4b040-0062-41bc-8b40-b269d0c9820b req-810703f8-2461-40e2-9f40-039847c3a180 service nova] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.530581] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.559092] env[62914]: DEBUG nova.network.neutron [req-03b4b040-0062-41bc-8b40-b269d0c9820b req-810703f8-2461-40e2-9f40-039847c3a180 service nova] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.563346] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352418, 'name': Rename_Task, 'duration_secs': 0.148856} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.563929] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 750.564045] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93fa221b-fc88-4b91-b857-5570c94cae0e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.570768] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 750.570768] env[62914]: value = "task-1352419" [ 750.570768] env[62914]: _type = "Task" [ 750.570768] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.583075] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352419, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.702327] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c1dfba-5851-4332-ad73-b3db4aea60e5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.710095] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9abaa68-ae8e-4a2e-96c5-ccfaa7eca322 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.742837] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8de2945-f7ef-478a-a408-2f3493e50961 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.750500] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c169bd89-3cf5-4afc-bb1a-fec49f41ffc3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.763781] env[62914]: DEBUG nova.compute.provider_tree [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.034896] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Releasing lock "refresh_cache-4c1016c4-2da8-4f4e-9497-4d680fb74535" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.034896] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 751.035365] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 751.035412] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 751.055549] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.064236] env[62914]: DEBUG oslo_concurrency.lockutils [req-03b4b040-0062-41bc-8b40-b269d0c9820b req-810703f8-2461-40e2-9f40-039847c3a180 service nova] Releasing lock "refresh_cache-98c19874-7c71-4e20-aa29-2ddc53df7e81" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.064740] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "refresh_cache-98c19874-7c71-4e20-aa29-2ddc53df7e81" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.065043] env[62914]: DEBUG nova.network.neutron [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.082347] env[62914]: DEBUG oslo_vmware.api [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352419, 'name': PowerOnVM_Task, 'duration_secs': 0.485831} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.082525] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 751.082718] env[62914]: INFO nova.compute.manager [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Took 5.35 seconds to spawn the instance on the hypervisor. [ 751.082891] env[62914]: DEBUG nova.compute.manager [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 751.083644] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0dc649f-88f8-44ea-bd55-3d6df91d2052 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.267576] env[62914]: DEBUG nova.scheduler.client.report [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.559108] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.584821] env[62914]: DEBUG nova.network.neutron [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.601219] env[62914]: INFO nova.compute.manager [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Took 25.81 seconds to build instance. [ 751.669188] env[62914]: DEBUG nova.network.neutron [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.772472] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.881s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.773273] env[62914]: ERROR nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 908ec2a7-3e08-4b06-8152-3c414c106cbb, please check neutron logs for more information. [ 751.773273] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Traceback (most recent call last): [ 751.773273] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 751.773273] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] self.driver.spawn(context, instance, image_meta, [ 751.773273] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 751.773273] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.773273] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.773273] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] vm_ref = self.build_virtual_machine(instance, [ 751.773273] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.773273] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.773273] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.773857] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] for vif in network_info: [ 751.773857] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 751.773857] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] return self._sync_wrapper(fn, *args, **kwargs) [ 751.773857] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 751.773857] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] self.wait() [ 751.773857] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 751.773857] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] self[:] = self._gt.wait() [ 751.773857] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.773857] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] return self._exit_event.wait() [ 751.773857] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 751.773857] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] result = hub.switch() [ 751.773857] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 751.773857] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] return self.greenlet.switch() [ 751.774524] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.774524] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] result = function(*args, **kwargs) [ 751.774524] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 751.774524] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] return func(*args, **kwargs) [ 751.774524] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 751.774524] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] raise e [ 751.774524] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.774524] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] nwinfo = self.network_api.allocate_for_instance( [ 751.774524] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.774524] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] created_port_ids = self._update_ports_for_instance( [ 751.774524] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.774524] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] with excutils.save_and_reraise_exception(): [ 751.774524] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.775192] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] self.force_reraise() [ 751.775192] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.775192] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] raise self.value [ 751.775192] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.775192] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] updated_port = self._update_port( [ 751.775192] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.775192] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] _ensure_no_port_binding_failure(port) [ 751.775192] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.775192] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] raise exception.PortBindingFailed(port_id=port['id']) [ 751.775192] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] nova.exception.PortBindingFailed: Binding failed for port 908ec2a7-3e08-4b06-8152-3c414c106cbb, please check neutron logs for more information. [ 751.775192] env[62914]: ERROR nova.compute.manager [instance: 41218b72-0b79-496a-855f-57a972dabd0b] [ 751.776203] env[62914]: DEBUG nova.compute.utils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Binding failed for port 908ec2a7-3e08-4b06-8152-3c414c106cbb, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 751.776203] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Build of instance 41218b72-0b79-496a-855f-57a972dabd0b was re-scheduled: Binding failed for port 908ec2a7-3e08-4b06-8152-3c414c106cbb, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 751.776203] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 751.776458] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "refresh_cache-41218b72-0b79-496a-855f-57a972dabd0b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.776458] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquired lock "refresh_cache-41218b72-0b79-496a-855f-57a972dabd0b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.776574] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.778183] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.098s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.980557] env[62914]: DEBUG nova.compute.manager [req-244962ae-9b9d-49f0-971a-8d186147ab91 req-bdf2c3cd-590a-4af8-a75a-36b47c97fe33 service nova] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Received event network-vif-deleted-f517bef4-915c-434b-95aa-baf973fd5446 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 752.064287] env[62914]: INFO nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 4c1016c4-2da8-4f4e-9497-4d680fb74535] Took 1.03 seconds to deallocate network for instance. [ 752.102159] env[62914]: DEBUG oslo_concurrency.lockutils [None req-30b2e932-1805-4b67-adcc-42ba42c61a6e tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Lock "44339b45-5183-4882-aaec-4070adc3c3dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.970s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.154032] env[62914]: INFO nova.compute.manager [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Rebuilding instance [ 752.173332] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "refresh_cache-98c19874-7c71-4e20-aa29-2ddc53df7e81" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.173768] env[62914]: DEBUG nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 752.173982] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 752.175838] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-757d71b6-ea8e-4cb8-9ad8-3e7646e0cf2d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.186071] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f960f3c7-62cd-4f7b-b3ff-b811bc2af67e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.201231] env[62914]: DEBUG nova.compute.manager [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 752.202135] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c0e2e4-b008-4c4a-98b0-75ccc2e675e3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.210032] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 98c19874-7c71-4e20-aa29-2ddc53df7e81 could not be found. [ 752.210193] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 752.210372] env[62914]: INFO nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Took 0.04 seconds to destroy the instance on the hypervisor. [ 752.210604] env[62914]: DEBUG oslo.service.loopingcall [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.212286] env[62914]: DEBUG nova.compute.manager [-] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.212286] env[62914]: DEBUG nova.network.neutron [-] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.230089] env[62914]: DEBUG nova.network.neutron [-] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.303074] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.359778] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "46d139a1-a4ae-435e-9d6f-cfc06d706128" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.360069] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "46d139a1-a4ae-435e-9d6f-cfc06d706128" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.397857] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.535734] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7556e6-1461-4765-902d-ef43a8ce50d2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.544825] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a06b06ca-e20d-4ece-a34a-29e50f7b4d21 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.582514] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2a4989-0c93-4bc6-892f-470cd1670cec {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.590011] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42faa6d6-1303-4cb4-b0f6-67487ee3b82f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.603969] env[62914]: DEBUG nova.compute.provider_tree [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.605294] env[62914]: DEBUG nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 752.717675] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 752.717943] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c7e6894-fe5d-4195-9b0c-b5cf79abc6de {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.724910] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 752.724910] env[62914]: value = "task-1352420" [ 752.724910] env[62914]: _type = "Task" [ 752.724910] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.732726] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352420, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.732975] env[62914]: DEBUG nova.network.neutron [-] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.900037] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Releasing lock "refresh_cache-41218b72-0b79-496a-855f-57a972dabd0b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.900270] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 752.900451] env[62914]: DEBUG nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.900632] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.915878] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.104570] env[62914]: INFO nova.scheduler.client.report [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Deleted allocations for instance 4c1016c4-2da8-4f4e-9497-4d680fb74535 [ 753.111616] env[62914]: DEBUG nova.scheduler.client.report [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 753.132468] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.234744] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352420, 'name': PowerOffVM_Task, 'duration_secs': 0.113686} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.235131] env[62914]: INFO nova.compute.manager [-] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Took 1.02 seconds to deallocate network for instance. [ 753.235433] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 753.235665] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 753.237754] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e44e8e4-a780-485f-9c00-fa58bb7a9dc1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.240471] env[62914]: DEBUG nova.compute.claims [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 753.240643] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.244800] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 753.245032] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-443bc148-8a5d-41af-b029-d84d9a2948c4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.269182] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 753.269392] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 753.269567] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Deleting the datastore file [datastore2] 44339b45-5183-4882-aaec-4070adc3c3dd {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 753.269837] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ac014a7d-e8b1-4336-ba7d-35bc0107afd1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.276829] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 753.276829] env[62914]: value = "task-1352422" [ 753.276829] env[62914]: _type = "Task" [ 753.276829] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.284622] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352422, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.418871] env[62914]: DEBUG nova.network.neutron [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.619062] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.841s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.619573] env[62914]: ERROR nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7c98c7f3-faf2-4254-b99e-3d2381fe8a24, please check neutron logs for more information. [ 753.619573] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Traceback (most recent call last): [ 753.619573] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 753.619573] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] self.driver.spawn(context, instance, image_meta, [ 753.619573] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 753.619573] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.619573] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.619573] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] vm_ref = self.build_virtual_machine(instance, [ 753.619573] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.619573] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.619573] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.619975] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] for vif in network_info: [ 753.619975] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.619975] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] return self._sync_wrapper(fn, *args, **kwargs) [ 753.619975] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.619975] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] self.wait() [ 753.619975] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.619975] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] self[:] = self._gt.wait() [ 753.619975] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.619975] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] return self._exit_event.wait() [ 753.619975] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 753.619975] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] current.throw(*self._exc) [ 753.619975] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.619975] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] result = function(*args, **kwargs) [ 753.620403] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 753.620403] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] return func(*args, **kwargs) [ 753.620403] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.620403] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] raise e [ 753.620403] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.620403] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] nwinfo = self.network_api.allocate_for_instance( [ 753.620403] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.620403] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] created_port_ids = self._update_ports_for_instance( [ 753.620403] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.620403] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] with excutils.save_and_reraise_exception(): [ 753.620403] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.620403] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] self.force_reraise() [ 753.620403] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.620880] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] raise self.value [ 753.620880] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.620880] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] updated_port = self._update_port( [ 753.620880] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.620880] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] _ensure_no_port_binding_failure(port) [ 753.620880] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.620880] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] raise exception.PortBindingFailed(port_id=port['id']) [ 753.620880] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] nova.exception.PortBindingFailed: Binding failed for port 7c98c7f3-faf2-4254-b99e-3d2381fe8a24, please check neutron logs for more information. [ 753.620880] env[62914]: ERROR nova.compute.manager [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] [ 753.620880] env[62914]: DEBUG nova.compute.utils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Binding failed for port 7c98c7f3-faf2-4254-b99e-3d2381fe8a24, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 753.622648] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.995s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.622819] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.622965] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62914) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 753.623285] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.120s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.627943] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "4c1016c4-2da8-4f4e-9497-4d680fb74535" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.900s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.628500] env[62914]: DEBUG nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Build of instance d9057a9e-988f-44ee-9339-ca9a1429378e was re-scheduled: Binding failed for port 7c98c7f3-faf2-4254-b99e-3d2381fe8a24, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 753.628912] env[62914]: DEBUG nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 753.629141] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Acquiring lock "refresh_cache-d9057a9e-988f-44ee-9339-ca9a1429378e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.629283] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Acquired lock "refresh_cache-d9057a9e-988f-44ee-9339-ca9a1429378e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.629445] env[62914]: DEBUG nova.network.neutron [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 753.630795] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c82434-3ce1-4d56-af20-aa8646929142 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.640154] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6fd4c2-292b-44ab-a41b-a183c5b05337 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.654379] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc9af42-2d49-491b-9ebc-fd1fdc0bf3c8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.661146] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c1a0d2-aa65-4bf0-b10d-17b9d340df84 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.689287] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181513MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62914) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 753.689420] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.786300] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352422, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101929} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.786537] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 753.786717] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 753.786886] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 753.921618] env[62914]: INFO nova.compute.manager [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 41218b72-0b79-496a-855f-57a972dabd0b] Took 1.02 seconds to deallocate network for instance. [ 754.137355] env[62914]: DEBUG nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 754.155809] env[62914]: DEBUG nova.network.neutron [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.232631] env[62914]: DEBUG nova.network.neutron [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.383308] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab0f59c-a131-4879-9d3e-082286e18d48 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.391464] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157cf1d4-5bfa-4175-8954-09cce4ea2b26 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.422969] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f854fb-e27c-413e-8fb9-f84b5063f4a5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.433715] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68f64db-0a75-4a1c-8a52-40dbc4a00bce {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.446601] env[62914]: DEBUG nova.compute.provider_tree [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.660843] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.735055] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Releasing lock "refresh_cache-d9057a9e-988f-44ee-9339-ca9a1429378e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.735282] env[62914]: DEBUG nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 754.735488] env[62914]: DEBUG nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.735651] env[62914]: DEBUG nova.network.neutron [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 754.750201] env[62914]: DEBUG nova.network.neutron [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.822043] env[62914]: DEBUG nova.virt.hardware [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 754.822348] env[62914]: DEBUG nova.virt.hardware [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 754.822540] env[62914]: DEBUG nova.virt.hardware [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 754.822758] env[62914]: DEBUG nova.virt.hardware [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 754.822930] env[62914]: DEBUG nova.virt.hardware [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 754.823121] env[62914]: DEBUG nova.virt.hardware [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 754.823380] env[62914]: DEBUG nova.virt.hardware [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 754.823586] env[62914]: DEBUG nova.virt.hardware [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 754.823764] env[62914]: DEBUG nova.virt.hardware [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 754.823922] env[62914]: DEBUG nova.virt.hardware [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 754.824108] env[62914]: DEBUG nova.virt.hardware [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 754.825211] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5088cb6c-f687-4145-b3f0-f7337dec0214 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.833677] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-147fa8f6-8458-48d4-a9db-853c4c1f04ce {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.848304] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Instance VIF info [] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 754.854146] env[62914]: DEBUG oslo.service.loopingcall [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.854250] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 754.854411] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c5370c31-07a9-4cdb-b43c-561bbe18f52e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.870799] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 754.870799] env[62914]: value = "task-1352423" [ 754.870799] env[62914]: _type = "Task" [ 754.870799] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.878276] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352423, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.950074] env[62914]: DEBUG nova.scheduler.client.report [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.954526] env[62914]: INFO nova.scheduler.client.report [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Deleted allocations for instance 41218b72-0b79-496a-855f-57a972dabd0b [ 755.253311] env[62914]: DEBUG nova.network.neutron [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.380776] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352423, 'name': CreateVM_Task, 'duration_secs': 0.251787} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.380925] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 755.381366] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.381691] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.381828] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 755.382086] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3056d659-1d95-4a94-abfc-56872fa7bc43 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.386255] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 755.386255] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]525d64eb-f4f5-67df-2397-b2c623809034" [ 755.386255] env[62914]: _type = "Task" [ 755.386255] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.393834] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]525d64eb-f4f5-67df-2397-b2c623809034, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.463569] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.840s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.463695] env[62914]: ERROR nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3345d945-7dfe-4801-b734-0c7a799d785a, please check neutron logs for more information. [ 755.463695] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Traceback (most recent call last): [ 755.463695] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 755.463695] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] self.driver.spawn(context, instance, image_meta, [ 755.463695] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 755.463695] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 755.463695] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 755.463695] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] vm_ref = self.build_virtual_machine(instance, [ 755.463695] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 755.463695] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] vif_infos = vmwarevif.get_vif_info(self._session, [ 755.463695] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 755.464317] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] for vif in network_info: [ 755.464317] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 755.464317] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] return self._sync_wrapper(fn, *args, **kwargs) [ 755.464317] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 755.464317] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] self.wait() [ 755.464317] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 755.464317] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] self[:] = self._gt.wait() [ 755.464317] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 755.464317] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] return self._exit_event.wait() [ 755.464317] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 755.464317] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] current.throw(*self._exc) [ 755.464317] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.464317] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] result = function(*args, **kwargs) [ 755.464759] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 755.464759] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] return func(*args, **kwargs) [ 755.464759] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 755.464759] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] raise e [ 755.464759] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.464759] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] nwinfo = self.network_api.allocate_for_instance( [ 755.464759] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 755.464759] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] created_port_ids = self._update_ports_for_instance( [ 755.464759] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 755.464759] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] with excutils.save_and_reraise_exception(): [ 755.464759] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.464759] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] self.force_reraise() [ 755.464759] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.465190] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] raise self.value [ 755.465190] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 755.465190] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] updated_port = self._update_port( [ 755.465190] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.465190] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] _ensure_no_port_binding_failure(port) [ 755.465190] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.465190] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] raise exception.PortBindingFailed(port_id=port['id']) [ 755.465190] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] nova.exception.PortBindingFailed: Binding failed for port 3345d945-7dfe-4801-b734-0c7a799d785a, please check neutron logs for more information. [ 755.465190] env[62914]: ERROR nova.compute.manager [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] [ 755.465190] env[62914]: DEBUG nova.compute.utils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Binding failed for port 3345d945-7dfe-4801-b734-0c7a799d785a, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 755.465888] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.668s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.469030] env[62914]: DEBUG nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Build of instance 220578cb-17b1-4b46-8c4a-f676c04fc5e8 was re-scheduled: Binding failed for port 3345d945-7dfe-4801-b734-0c7a799d785a, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 755.469340] env[62914]: DEBUG nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 755.469566] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Acquiring lock "refresh_cache-220578cb-17b1-4b46-8c4a-f676c04fc5e8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.469736] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Acquired lock "refresh_cache-220578cb-17b1-4b46-8c4a-f676c04fc5e8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.469870] env[62914]: DEBUG nova.network.neutron [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 755.470922] env[62914]: DEBUG oslo_concurrency.lockutils [None req-59825a00-c51c-468f-a9aa-d928dc614cde tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "41218b72-0b79-496a-855f-57a972dabd0b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.705s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.755459] env[62914]: INFO nova.compute.manager [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: d9057a9e-988f-44ee-9339-ca9a1429378e] Took 1.02 seconds to deallocate network for instance. [ 755.895950] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]525d64eb-f4f5-67df-2397-b2c623809034, 'name': SearchDatastore_Task, 'duration_secs': 0.009197} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.896140] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.896363] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 755.896588] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.896728] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.896895] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 755.897176] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6fa3327c-7406-45fe-91ff-68a89880283e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.906025] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 755.906025] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 755.906025] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f292fed6-22ef-4819-bd07-81e8afd672a5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.910150] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 755.910150] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c7df14-ce2a-1d1b-d9bb-8380d8c525eb" [ 755.910150] env[62914]: _type = "Task" [ 755.910150] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.917851] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c7df14-ce2a-1d1b-d9bb-8380d8c525eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.976754] env[62914]: DEBUG nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 755.999734] env[62914]: DEBUG nova.network.neutron [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.094416] env[62914]: DEBUG nova.network.neutron [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.215213] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20daf670-0ede-4a0b-98f8-3fe15a161295 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.223148] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1644fb32-be0b-4553-8d7f-2d724cac2d4b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.254041] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e864c9bb-3123-4637-89f6-6d80777cc840 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.264049] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f61d80-22b2-42ad-b8e4-491c462349fb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.277361] env[62914]: DEBUG nova.compute.provider_tree [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.420413] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c7df14-ce2a-1d1b-d9bb-8380d8c525eb, 'name': SearchDatastore_Task, 'duration_secs': 0.007775} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.421191] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b84d3a0b-335f-4b86-85f8-d7c57c760460 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.426467] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 756.426467] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5227b944-5d7e-9b03-3129-b192247c8137" [ 756.426467] env[62914]: _type = "Task" [ 756.426467] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.433989] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5227b944-5d7e-9b03-3129-b192247c8137, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.496979] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.514497] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "cb89c84f-414b-4dc7-9db5-12cbc30b52b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.514759] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "cb89c84f-414b-4dc7-9db5-12cbc30b52b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.536751] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "9577394d-1c73-4ed1-ba86-e7c246e32719" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.536953] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "9577394d-1c73-4ed1-ba86-e7c246e32719" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.596668] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Releasing lock "refresh_cache-220578cb-17b1-4b46-8c4a-f676c04fc5e8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.596896] env[62914]: DEBUG nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 756.597088] env[62914]: DEBUG nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 756.597284] env[62914]: DEBUG nova.network.neutron [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 756.613781] env[62914]: DEBUG nova.network.neutron [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.783527] env[62914]: DEBUG nova.scheduler.client.report [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 756.795217] env[62914]: INFO nova.scheduler.client.report [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Deleted allocations for instance d9057a9e-988f-44ee-9339-ca9a1429378e [ 756.939010] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5227b944-5d7e-9b03-3129-b192247c8137, 'name': SearchDatastore_Task, 'duration_secs': 0.009338} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.939282] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.939547] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 44339b45-5183-4882-aaec-4070adc3c3dd/44339b45-5183-4882-aaec-4070adc3c3dd.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 756.939789] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7cb827da-e83b-46c9-a2ce-3caeeae90f37 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.946345] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 756.946345] env[62914]: value = "task-1352424" [ 756.946345] env[62914]: _type = "Task" [ 756.946345] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.953931] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352424, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.116810] env[62914]: DEBUG nova.network.neutron [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.288824] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.823s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.289554] env[62914]: ERROR nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 385e4c37-f5d7-4035-b356-34fe317498c4, please check neutron logs for more information. [ 757.289554] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Traceback (most recent call last): [ 757.289554] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 757.289554] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] self.driver.spawn(context, instance, image_meta, [ 757.289554] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 757.289554] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.289554] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.289554] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] vm_ref = self.build_virtual_machine(instance, [ 757.289554] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.289554] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.289554] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.289950] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] for vif in network_info: [ 757.289950] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.289950] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] return self._sync_wrapper(fn, *args, **kwargs) [ 757.289950] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.289950] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] self.wait() [ 757.289950] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.289950] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] self[:] = self._gt.wait() [ 757.289950] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.289950] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] return self._exit_event.wait() [ 757.289950] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 757.289950] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] current.throw(*self._exc) [ 757.289950] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.289950] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] result = function(*args, **kwargs) [ 757.290325] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 757.290325] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] return func(*args, **kwargs) [ 757.290325] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.290325] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] raise e [ 757.290325] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.290325] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] nwinfo = self.network_api.allocate_for_instance( [ 757.290325] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.290325] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] created_port_ids = self._update_ports_for_instance( [ 757.290325] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.290325] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] with excutils.save_and_reraise_exception(): [ 757.290325] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.290325] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] self.force_reraise() [ 757.290325] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.290742] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] raise self.value [ 757.290742] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.290742] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] updated_port = self._update_port( [ 757.290742] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.290742] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] _ensure_no_port_binding_failure(port) [ 757.290742] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.290742] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] raise exception.PortBindingFailed(port_id=port['id']) [ 757.290742] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] nova.exception.PortBindingFailed: Binding failed for port 385e4c37-f5d7-4035-b356-34fe317498c4, please check neutron logs for more information. [ 757.290742] env[62914]: ERROR nova.compute.manager [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] [ 757.290742] env[62914]: DEBUG nova.compute.utils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Binding failed for port 385e4c37-f5d7-4035-b356-34fe317498c4, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 757.292179] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.086s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.293941] env[62914]: INFO nova.compute.claims [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.297123] env[62914]: DEBUG nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Build of instance d73f5ea8-529a-44e3-8239-5aaa4534d1cc was re-scheduled: Binding failed for port 385e4c37-f5d7-4035-b356-34fe317498c4, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 757.297520] env[62914]: DEBUG nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 757.297768] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Acquiring lock "refresh_cache-d73f5ea8-529a-44e3-8239-5aaa4534d1cc" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.297899] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Acquired lock "refresh_cache-d73f5ea8-529a-44e3-8239-5aaa4534d1cc" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.298068] env[62914]: DEBUG nova.network.neutron [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.306393] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7675293e-593b-4d9b-aab1-bd0bbf0b2911 tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Lock "d9057a9e-988f-44ee-9339-ca9a1429378e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.078s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.457293] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352424, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470372} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.457682] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 44339b45-5183-4882-aaec-4070adc3c3dd/44339b45-5183-4882-aaec-4070adc3c3dd.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 757.458014] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 757.458368] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5a93ef7a-a9e8-4b8a-bb66-0b63ea828115 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.464807] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 757.464807] env[62914]: value = "task-1352425" [ 757.464807] env[62914]: _type = "Task" [ 757.464807] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.472570] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352425, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.619368] env[62914]: INFO nova.compute.manager [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] [instance: 220578cb-17b1-4b46-8c4a-f676c04fc5e8] Took 1.02 seconds to deallocate network for instance. [ 757.808633] env[62914]: DEBUG nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.817581] env[62914]: DEBUG nova.network.neutron [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.870788] env[62914]: DEBUG nova.network.neutron [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.975020] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352425, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061761} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.975145] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 757.975840] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377a3616-56d6-4bd8-ad7f-d30e8ae5acf8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.995491] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 44339b45-5183-4882-aaec-4070adc3c3dd/44339b45-5183-4882-aaec-4070adc3c3dd.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 757.995793] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ae0158f-92f7-4458-92cf-b211174442e5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.015869] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 758.015869] env[62914]: value = "task-1352426" [ 758.015869] env[62914]: _type = "Task" [ 758.015869] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.023381] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352426, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.328941] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.373483] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Releasing lock "refresh_cache-d73f5ea8-529a-44e3-8239-5aaa4534d1cc" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.373990] env[62914]: DEBUG nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 758.373990] env[62914]: DEBUG nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.374156] env[62914]: DEBUG nova.network.neutron [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 758.388633] env[62914]: DEBUG nova.network.neutron [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.527962] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352426, 'name': ReconfigVM_Task, 'duration_secs': 0.276816} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.529021] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 44339b45-5183-4882-aaec-4070adc3c3dd/44339b45-5183-4882-aaec-4070adc3c3dd.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 758.530080] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea2bade5-1707-4ac1-840a-9f6afdc2765a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.532464] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa9a8ff8-565e-47ed-beb5-90c2ce2c76a4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.539053] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efde09ac-7b54-4e23-983b-8194d9877406 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.541961] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 758.541961] env[62914]: value = "task-1352427" [ 758.541961] env[62914]: _type = "Task" [ 758.541961] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.571886] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d89e3e-7415-4dfd-abff-fee1cace5ba8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.574381] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352427, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.579050] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64969fb2-9a21-49ae-978a-4479187cc5a1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.591724] env[62914]: DEBUG nova.compute.provider_tree [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.654991] env[62914]: INFO nova.scheduler.client.report [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Deleted allocations for instance 220578cb-17b1-4b46-8c4a-f676c04fc5e8 [ 758.892024] env[62914]: DEBUG nova.network.neutron [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.051809] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352427, 'name': Rename_Task, 'duration_secs': 0.134175} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.052100] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 759.052345] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4fcbdb85-81d6-4656-989c-4e2f9a334d36 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.058529] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 759.058529] env[62914]: value = "task-1352428" [ 759.058529] env[62914]: _type = "Task" [ 759.058529] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.065839] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352428, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.095024] env[62914]: DEBUG nova.scheduler.client.report [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.165280] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb332ba3-02d9-44e1-ad09-3ebaf056776f tempest-ServersAdminTestJSON-1767833814 tempest-ServersAdminTestJSON-1767833814-project-member] Lock "220578cb-17b1-4b46-8c4a-f676c04fc5e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.322s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.395935] env[62914]: INFO nova.compute.manager [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] [instance: d73f5ea8-529a-44e3-8239-5aaa4534d1cc] Took 1.02 seconds to deallocate network for instance. [ 759.573014] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352428, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.600104] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.600706] env[62914]: DEBUG nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 759.604272] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.549s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.606553] env[62914]: INFO nova.compute.claims [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.669034] env[62914]: DEBUG nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 760.069968] env[62914]: DEBUG oslo_vmware.api [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352428, 'name': PowerOnVM_Task, 'duration_secs': 1.005018} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.070452] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 760.070753] env[62914]: DEBUG nova.compute.manager [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 760.071648] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b53c1b7b-75cb-442e-8c05-d5ec2388da0c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.111778] env[62914]: DEBUG nova.compute.utils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.115373] env[62914]: DEBUG nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 760.115630] env[62914]: DEBUG nova.network.neutron [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 760.191017] env[62914]: DEBUG nova.policy [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd900131bc974504a489fcd4ffaec5b6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c3055903b6f4443bbeb9897e75e34c49', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 760.199427] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.427315] env[62914]: INFO nova.scheduler.client.report [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Deleted allocations for instance d73f5ea8-529a-44e3-8239-5aaa4534d1cc [ 760.591437] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.619290] env[62914]: DEBUG nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 760.692515] env[62914]: DEBUG nova.network.neutron [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Successfully created port: ea8ae277-71c8-4d99-b08e-9c135e326363 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 760.908148] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a1f6d92-73cd-4b47-8d3d-1f350ffbdf0a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.916018] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24911232-2272-4d52-87fe-5550d23308a2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.944908] env[62914]: DEBUG oslo_concurrency.lockutils [None req-31d6bc9d-752f-4e2a-a4b4-9c7af591f12a tempest-TenantUsagesTestJSON-885592675 tempest-TenantUsagesTestJSON-885592675-project-member] Lock "d73f5ea8-529a-44e3-8239-5aaa4534d1cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.804s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.950995] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b349ff-46dc-477a-b1fb-774a4985b13d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.955858] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e937381-b181-4f00-b78a-7c667cad4c15 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.963019] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquiring lock "44339b45-5183-4882-aaec-4070adc3c3dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.963019] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Lock "44339b45-5183-4882-aaec-4070adc3c3dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.963019] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquiring lock "44339b45-5183-4882-aaec-4070adc3c3dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.963019] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Lock "44339b45-5183-4882-aaec-4070adc3c3dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.963413] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Lock "44339b45-5183-4882-aaec-4070adc3c3dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.964193] env[62914]: INFO nova.compute.manager [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Terminating instance [ 760.973524] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquiring lock "refresh_cache-44339b45-5183-4882-aaec-4070adc3c3dd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.973853] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquired lock "refresh_cache-44339b45-5183-4882-aaec-4070adc3c3dd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.976610] env[62914]: DEBUG nova.network.neutron [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 760.976610] env[62914]: DEBUG nova.compute.provider_tree [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.452104] env[62914]: DEBUG nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 761.480550] env[62914]: DEBUG nova.scheduler.client.report [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.507021] env[62914]: DEBUG nova.network.neutron [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.609017] env[62914]: DEBUG nova.network.neutron [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.695120] env[62914]: DEBUG nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 761.748020] env[62914]: DEBUG nova.virt.hardware [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 761.748020] env[62914]: DEBUG nova.virt.hardware [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 761.748020] env[62914]: DEBUG nova.virt.hardware [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 761.748301] env[62914]: DEBUG nova.virt.hardware [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 761.748301] env[62914]: DEBUG nova.virt.hardware [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 761.748301] env[62914]: DEBUG nova.virt.hardware [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 761.748301] env[62914]: DEBUG nova.virt.hardware [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 761.748301] env[62914]: DEBUG nova.virt.hardware [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 761.748508] env[62914]: DEBUG nova.virt.hardware [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 761.748508] env[62914]: DEBUG nova.virt.hardware [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 761.748508] env[62914]: DEBUG nova.virt.hardware [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.748508] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e225d9-286d-4247-970a-e0ce11d00863 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.756564] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f252361-7ee0-4dd6-94a2-ef45a792e383 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.977492] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.989409] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.993019] env[62914]: DEBUG nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 761.993798] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.657s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.000020] env[62914]: INFO nova.compute.claims [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.114119] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Releasing lock "refresh_cache-44339b45-5183-4882-aaec-4070adc3c3dd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.114544] env[62914]: DEBUG nova.compute.manager [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 762.114735] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 762.115674] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343fdffa-9f54-4ad7-8731-6e4fcc0d16b5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.122808] env[62914]: DEBUG nova.compute.manager [req-c709cc58-1748-4013-bed7-24c9c3df2574 req-5c006ae7-34e8-4b95-987c-17f225715577 service nova] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Received event network-changed-ea8ae277-71c8-4d99-b08e-9c135e326363 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.122982] env[62914]: DEBUG nova.compute.manager [req-c709cc58-1748-4013-bed7-24c9c3df2574 req-5c006ae7-34e8-4b95-987c-17f225715577 service nova] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Refreshing instance network info cache due to event network-changed-ea8ae277-71c8-4d99-b08e-9c135e326363. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 762.123182] env[62914]: DEBUG oslo_concurrency.lockutils [req-c709cc58-1748-4013-bed7-24c9c3df2574 req-5c006ae7-34e8-4b95-987c-17f225715577 service nova] Acquiring lock "refresh_cache-364c5d87-f261-439a-920e-874ee0bcf83b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.123330] env[62914]: DEBUG oslo_concurrency.lockutils [req-c709cc58-1748-4013-bed7-24c9c3df2574 req-5c006ae7-34e8-4b95-987c-17f225715577 service nova] Acquired lock "refresh_cache-364c5d87-f261-439a-920e-874ee0bcf83b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.123456] env[62914]: DEBUG nova.network.neutron [req-c709cc58-1748-4013-bed7-24c9c3df2574 req-5c006ae7-34e8-4b95-987c-17f225715577 service nova] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Refreshing network info cache for port ea8ae277-71c8-4d99-b08e-9c135e326363 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 762.126780] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 762.127213] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25db3253-b62c-4883-b8d4-83cc96c928c1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.133798] env[62914]: DEBUG oslo_vmware.api [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 762.133798] env[62914]: value = "task-1352429" [ 762.133798] env[62914]: _type = "Task" [ 762.133798] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.142303] env[62914]: DEBUG oslo_vmware.api [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352429, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.464132] env[62914]: ERROR nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ea8ae277-71c8-4d99-b08e-9c135e326363, please check neutron logs for more information. [ 762.464132] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 762.464132] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.464132] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 762.464132] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.464132] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 762.464132] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.464132] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 762.464132] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.464132] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 762.464132] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.464132] env[62914]: ERROR nova.compute.manager raise self.value [ 762.464132] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.464132] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 762.464132] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.464132] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 762.464956] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.464956] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 762.464956] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ea8ae277-71c8-4d99-b08e-9c135e326363, please check neutron logs for more information. [ 762.464956] env[62914]: ERROR nova.compute.manager [ 762.464956] env[62914]: Traceback (most recent call last): [ 762.464956] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 762.464956] env[62914]: listener.cb(fileno) [ 762.464956] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.464956] env[62914]: result = function(*args, **kwargs) [ 762.464956] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 762.464956] env[62914]: return func(*args, **kwargs) [ 762.464956] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.464956] env[62914]: raise e [ 762.464956] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.464956] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 762.464956] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.464956] env[62914]: created_port_ids = self._update_ports_for_instance( [ 762.464956] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.464956] env[62914]: with excutils.save_and_reraise_exception(): [ 762.464956] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.464956] env[62914]: self.force_reraise() [ 762.464956] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.464956] env[62914]: raise self.value [ 762.464956] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.464956] env[62914]: updated_port = self._update_port( [ 762.464956] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.464956] env[62914]: _ensure_no_port_binding_failure(port) [ 762.464956] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.464956] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 762.465846] env[62914]: nova.exception.PortBindingFailed: Binding failed for port ea8ae277-71c8-4d99-b08e-9c135e326363, please check neutron logs for more information. [ 762.465846] env[62914]: Removing descriptor: 19 [ 762.465846] env[62914]: ERROR nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ea8ae277-71c8-4d99-b08e-9c135e326363, please check neutron logs for more information. [ 762.465846] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Traceback (most recent call last): [ 762.465846] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 762.465846] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] yield resources [ 762.465846] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 762.465846] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] self.driver.spawn(context, instance, image_meta, [ 762.465846] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 762.465846] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.465846] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.465846] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] vm_ref = self.build_virtual_machine(instance, [ 762.466223] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.466223] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.466223] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.466223] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] for vif in network_info: [ 762.466223] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 762.466223] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] return self._sync_wrapper(fn, *args, **kwargs) [ 762.466223] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 762.466223] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] self.wait() [ 762.466223] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 762.466223] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] self[:] = self._gt.wait() [ 762.466223] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.466223] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] return self._exit_event.wait() [ 762.466223] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 762.466615] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] result = hub.switch() [ 762.466615] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 762.466615] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] return self.greenlet.switch() [ 762.466615] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.466615] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] result = function(*args, **kwargs) [ 762.466615] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 762.466615] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] return func(*args, **kwargs) [ 762.466615] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.466615] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] raise e [ 762.466615] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.466615] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] nwinfo = self.network_api.allocate_for_instance( [ 762.466615] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.466615] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] created_port_ids = self._update_ports_for_instance( [ 762.467102] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.467102] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] with excutils.save_and_reraise_exception(): [ 762.467102] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.467102] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] self.force_reraise() [ 762.467102] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.467102] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] raise self.value [ 762.467102] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.467102] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] updated_port = self._update_port( [ 762.467102] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.467102] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] _ensure_no_port_binding_failure(port) [ 762.467102] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.467102] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] raise exception.PortBindingFailed(port_id=port['id']) [ 762.467454] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] nova.exception.PortBindingFailed: Binding failed for port ea8ae277-71c8-4d99-b08e-9c135e326363, please check neutron logs for more information. [ 762.467454] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] [ 762.467454] env[62914]: INFO nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Terminating instance [ 762.470181] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "refresh_cache-364c5d87-f261-439a-920e-874ee0bcf83b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.499021] env[62914]: DEBUG nova.compute.utils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 762.499244] env[62914]: DEBUG nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 762.499621] env[62914]: DEBUG nova.network.neutron [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 762.596130] env[62914]: DEBUG nova.policy [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'baf190d4a357468a9eb18dfad17e3da0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '15315f8e77d749a4b2b0211c3b22b546', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 762.650415] env[62914]: DEBUG oslo_vmware.api [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352429, 'name': PowerOffVM_Task, 'duration_secs': 0.241159} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.651491] env[62914]: DEBUG nova.network.neutron [req-c709cc58-1748-4013-bed7-24c9c3df2574 req-5c006ae7-34e8-4b95-987c-17f225715577 service nova] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.653642] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 762.654095] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 762.654496] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fcbc8397-6ea9-4389-a5b7-8628cda79feb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.684024] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 762.684024] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 762.684024] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Deleting the datastore file [datastore2] 44339b45-5183-4882-aaec-4070adc3c3dd {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 762.684024] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be69e604-abe6-4f65-98e4-7d971d9e1c5d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.692100] env[62914]: DEBUG oslo_vmware.api [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for the task: (returnval){ [ 762.692100] env[62914]: value = "task-1352431" [ 762.692100] env[62914]: _type = "Task" [ 762.692100] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.698772] env[62914]: DEBUG oslo_vmware.api [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352431, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.748473] env[62914]: DEBUG nova.network.neutron [req-c709cc58-1748-4013-bed7-24c9c3df2574 req-5c006ae7-34e8-4b95-987c-17f225715577 service nova] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.006162] env[62914]: DEBUG nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 763.200283] env[62914]: DEBUG oslo_vmware.api [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Task: {'id': task-1352431, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140172} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.203087] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 763.203303] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 763.203503] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 763.203676] env[62914]: INFO nova.compute.manager [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Took 1.09 seconds to destroy the instance on the hypervisor. [ 763.203908] env[62914]: DEBUG oslo.service.loopingcall [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.204705] env[62914]: DEBUG nova.compute.manager [-] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 763.204805] env[62914]: DEBUG nova.network.neutron [-] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 763.223483] env[62914]: DEBUG nova.network.neutron [-] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.249756] env[62914]: DEBUG oslo_concurrency.lockutils [req-c709cc58-1748-4013-bed7-24c9c3df2574 req-5c006ae7-34e8-4b95-987c-17f225715577 service nova] Releasing lock "refresh_cache-364c5d87-f261-439a-920e-874ee0bcf83b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.250147] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquired lock "refresh_cache-364c5d87-f261-439a-920e-874ee0bcf83b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.250322] env[62914]: DEBUG nova.network.neutron [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.282240] env[62914]: DEBUG nova.network.neutron [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Successfully created port: 6e9fb149-2aaa-4699-ac9d-371d56feba58 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 763.308024] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a72431e-d75f-4295-906f-747a1a938587 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.313940] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b6718f-dcce-4604-9ffa-8d40d0ab10ea {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.347794] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3a273b-7180-4689-bd42-2594fa3370ee {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.357122] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8dd66f4-5722-4231-87ad-2feb62bac075 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.370938] env[62914]: DEBUG nova.compute.provider_tree [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.727296] env[62914]: DEBUG nova.network.neutron [-] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.775287] env[62914]: DEBUG nova.network.neutron [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.867160] env[62914]: DEBUG nova.network.neutron [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.879191] env[62914]: DEBUG nova.scheduler.client.report [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.015631] env[62914]: DEBUG nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 764.041909] env[62914]: DEBUG nova.virt.hardware [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 764.042160] env[62914]: DEBUG nova.virt.hardware [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 764.042311] env[62914]: DEBUG nova.virt.hardware [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 764.042482] env[62914]: DEBUG nova.virt.hardware [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 764.042762] env[62914]: DEBUG nova.virt.hardware [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 764.042762] env[62914]: DEBUG nova.virt.hardware [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 764.043062] env[62914]: DEBUG nova.virt.hardware [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 764.043357] env[62914]: DEBUG nova.virt.hardware [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 764.043567] env[62914]: DEBUG nova.virt.hardware [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 764.043732] env[62914]: DEBUG nova.virt.hardware [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 764.043896] env[62914]: DEBUG nova.virt.hardware [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 764.044931] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb902e7-ec17-4ec8-8f13-a9dc0d3ce330 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.054183] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355d5520-0b94-4a09-95fe-2534667b9790 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.213335] env[62914]: DEBUG nova.compute.manager [req-ad6e5887-a1f4-458e-8057-fb41bcdda8bd req-4065c1ed-d6e7-475e-af29-2b722b9511c4 service nova] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Received event network-vif-deleted-ea8ae277-71c8-4d99-b08e-9c135e326363 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.229257] env[62914]: INFO nova.compute.manager [-] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Took 1.02 seconds to deallocate network for instance. [ 764.368901] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Releasing lock "refresh_cache-364c5d87-f261-439a-920e-874ee0bcf83b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.369357] env[62914]: DEBUG nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 764.369520] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 764.369867] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-613f578b-dc98-46ad-920f-c385da0a21f1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.382494] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b76be9b3-3788-4122-9f33-194d493caa92 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.400137] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.400859] env[62914]: DEBUG nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 764.404569] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.271s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.404569] env[62914]: INFO nova.compute.claims [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 764.423446] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 364c5d87-f261-439a-920e-874ee0bcf83b could not be found. [ 764.423593] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 764.424417] env[62914]: INFO nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 764.424417] env[62914]: DEBUG oslo.service.loopingcall [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.424417] env[62914]: DEBUG nova.compute.manager [-] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.424417] env[62914]: DEBUG nova.network.neutron [-] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 764.451549] env[62914]: DEBUG nova.network.neutron [-] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.515906] env[62914]: ERROR nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6e9fb149-2aaa-4699-ac9d-371d56feba58, please check neutron logs for more information. [ 764.515906] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 764.515906] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.515906] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 764.515906] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.515906] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 764.515906] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.515906] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 764.515906] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.515906] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 764.515906] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.515906] env[62914]: ERROR nova.compute.manager raise self.value [ 764.515906] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.515906] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 764.515906] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.515906] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 764.516392] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.516392] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 764.516392] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6e9fb149-2aaa-4699-ac9d-371d56feba58, please check neutron logs for more information. [ 764.516392] env[62914]: ERROR nova.compute.manager [ 764.516392] env[62914]: Traceback (most recent call last): [ 764.516392] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 764.516392] env[62914]: listener.cb(fileno) [ 764.516392] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.516392] env[62914]: result = function(*args, **kwargs) [ 764.516392] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.516392] env[62914]: return func(*args, **kwargs) [ 764.516392] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.516392] env[62914]: raise e [ 764.516392] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.516392] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 764.516392] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.516392] env[62914]: created_port_ids = self._update_ports_for_instance( [ 764.516392] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.516392] env[62914]: with excutils.save_and_reraise_exception(): [ 764.516392] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.516392] env[62914]: self.force_reraise() [ 764.516392] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.516392] env[62914]: raise self.value [ 764.516392] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.516392] env[62914]: updated_port = self._update_port( [ 764.516392] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.516392] env[62914]: _ensure_no_port_binding_failure(port) [ 764.516392] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.516392] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 764.517201] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 6e9fb149-2aaa-4699-ac9d-371d56feba58, please check neutron logs for more information. [ 764.517201] env[62914]: Removing descriptor: 19 [ 764.517201] env[62914]: ERROR nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6e9fb149-2aaa-4699-ac9d-371d56feba58, please check neutron logs for more information. [ 764.517201] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] Traceback (most recent call last): [ 764.517201] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 764.517201] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] yield resources [ 764.517201] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 764.517201] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] self.driver.spawn(context, instance, image_meta, [ 764.517201] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 764.517201] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 764.517201] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 764.517201] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] vm_ref = self.build_virtual_machine(instance, [ 764.517613] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 764.517613] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] vif_infos = vmwarevif.get_vif_info(self._session, [ 764.517613] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 764.517613] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] for vif in network_info: [ 764.517613] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 764.517613] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] return self._sync_wrapper(fn, *args, **kwargs) [ 764.517613] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 764.517613] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] self.wait() [ 764.517613] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 764.517613] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] self[:] = self._gt.wait() [ 764.517613] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 764.517613] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] return self._exit_event.wait() [ 764.517613] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 764.518035] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] result = hub.switch() [ 764.518035] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 764.518035] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] return self.greenlet.switch() [ 764.518035] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.518035] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] result = function(*args, **kwargs) [ 764.518035] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.518035] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] return func(*args, **kwargs) [ 764.518035] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.518035] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] raise e [ 764.518035] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.518035] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] nwinfo = self.network_api.allocate_for_instance( [ 764.518035] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.518035] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] created_port_ids = self._update_ports_for_instance( [ 764.518411] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.518411] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] with excutils.save_and_reraise_exception(): [ 764.518411] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.518411] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] self.force_reraise() [ 764.518411] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.518411] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] raise self.value [ 764.518411] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.518411] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] updated_port = self._update_port( [ 764.518411] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.518411] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] _ensure_no_port_binding_failure(port) [ 764.518411] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.518411] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] raise exception.PortBindingFailed(port_id=port['id']) [ 764.518816] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] nova.exception.PortBindingFailed: Binding failed for port 6e9fb149-2aaa-4699-ac9d-371d56feba58, please check neutron logs for more information. [ 764.518816] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] [ 764.518816] env[62914]: INFO nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Terminating instance [ 764.522329] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "refresh_cache-39489032-5430-471c-8d44-2c74682ed03b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.522932] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquired lock "refresh_cache-39489032-5430-471c-8d44-2c74682ed03b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.522932] env[62914]: DEBUG nova.network.neutron [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 764.740018] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.910906] env[62914]: DEBUG nova.compute.utils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 764.913410] env[62914]: DEBUG nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 764.913496] env[62914]: DEBUG nova.network.neutron [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 764.955374] env[62914]: DEBUG nova.network.neutron [-] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.962726] env[62914]: DEBUG nova.policy [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6dcc35fb57b46a2a753d8e4319ff290', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08363f37cd4246ba97abde2dbb87bf8b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.051114] env[62914]: DEBUG nova.network.neutron [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.192536] env[62914]: DEBUG nova.network.neutron [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.417426] env[62914]: DEBUG nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 765.457825] env[62914]: INFO nova.compute.manager [-] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Took 1.03 seconds to deallocate network for instance. [ 765.462931] env[62914]: DEBUG nova.compute.claims [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 765.463129] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.463877] env[62914]: DEBUG nova.network.neutron [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Successfully created port: d34e9c7a-0dd0-4619-9d18-9fc611717c55 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.697949] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Releasing lock "refresh_cache-39489032-5430-471c-8d44-2c74682ed03b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.698362] env[62914]: DEBUG nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 765.698552] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 765.700549] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-80619e4f-3125-466d-a421-5df9849c4d85 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.708913] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fde3099-b5e3-4364-892d-04c4d2c2dc50 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.738065] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 39489032-5430-471c-8d44-2c74682ed03b could not be found. [ 765.738263] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 765.738507] env[62914]: INFO nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 765.738690] env[62914]: DEBUG oslo.service.loopingcall [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.738972] env[62914]: DEBUG nova.compute.manager [-] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 765.739072] env[62914]: DEBUG nova.network.neutron [-] [instance: 39489032-5430-471c-8d44-2c74682ed03b] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 765.766621] env[62914]: DEBUG nova.network.neutron [-] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.793340] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a3aa28-0c40-4e5d-9f9e-edbdeae25704 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.801505] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d57fcd-2e93-4777-89b9-ffac8e19338b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.837305] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18f3466-7f98-426d-8b9f-56049daede26 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.845109] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d729864d-ba07-4b83-9bc2-f36de0c2dafc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.863912] env[62914]: DEBUG nova.compute.provider_tree [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.272768] env[62914]: DEBUG nova.network.neutron [-] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.366920] env[62914]: DEBUG nova.scheduler.client.report [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.423121] env[62914]: DEBUG nova.compute.manager [req-aa3a94f5-8a1f-4333-9fcf-b219d31769d4 req-2d367e74-123a-4124-8ac0-14099b4e4d38 service nova] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Received event network-changed-6e9fb149-2aaa-4699-ac9d-371d56feba58 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.423313] env[62914]: DEBUG nova.compute.manager [req-aa3a94f5-8a1f-4333-9fcf-b219d31769d4 req-2d367e74-123a-4124-8ac0-14099b4e4d38 service nova] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Refreshing instance network info cache due to event network-changed-6e9fb149-2aaa-4699-ac9d-371d56feba58. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 766.424090] env[62914]: DEBUG oslo_concurrency.lockutils [req-aa3a94f5-8a1f-4333-9fcf-b219d31769d4 req-2d367e74-123a-4124-8ac0-14099b4e4d38 service nova] Acquiring lock "refresh_cache-39489032-5430-471c-8d44-2c74682ed03b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.425994] env[62914]: DEBUG oslo_concurrency.lockutils [req-aa3a94f5-8a1f-4333-9fcf-b219d31769d4 req-2d367e74-123a-4124-8ac0-14099b4e4d38 service nova] Acquired lock "refresh_cache-39489032-5430-471c-8d44-2c74682ed03b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.426288] env[62914]: DEBUG nova.network.neutron [req-aa3a94f5-8a1f-4333-9fcf-b219d31769d4 req-2d367e74-123a-4124-8ac0-14099b4e4d38 service nova] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Refreshing network info cache for port 6e9fb149-2aaa-4699-ac9d-371d56feba58 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 766.436915] env[62914]: DEBUG nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 766.465602] env[62914]: DEBUG nova.virt.hardware [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 766.466266] env[62914]: DEBUG nova.virt.hardware [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 766.466266] env[62914]: DEBUG nova.virt.hardware [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.466266] env[62914]: DEBUG nova.virt.hardware [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 766.466532] env[62914]: DEBUG nova.virt.hardware [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.466532] env[62914]: DEBUG nova.virt.hardware [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 766.467042] env[62914]: DEBUG nova.virt.hardware [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 766.467234] env[62914]: DEBUG nova.virt.hardware [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 766.467671] env[62914]: DEBUG nova.virt.hardware [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 766.467906] env[62914]: DEBUG nova.virt.hardware [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 766.468105] env[62914]: DEBUG nova.virt.hardware [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.469607] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6447421a-90e0-4bd7-ad5e-39461fdb32f0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.482925] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f7dc3f-1ec5-4ca3-8c62-601c3bcdcc2a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.777580] env[62914]: INFO nova.compute.manager [-] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Took 1.04 seconds to deallocate network for instance. [ 766.780139] env[62914]: DEBUG nova.compute.claims [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 766.780335] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.879379] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.879379] env[62914]: DEBUG nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 766.881914] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.638s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.955631] env[62914]: DEBUG nova.network.neutron [req-aa3a94f5-8a1f-4333-9fcf-b219d31769d4 req-2d367e74-123a-4124-8ac0-14099b4e4d38 service nova] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.999925] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Acquiring lock "93d420a1-6d8f-4919-a42f-55aebab853ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.001310] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lock "93d420a1-6d8f-4919-a42f-55aebab853ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.039924] env[62914]: DEBUG nova.network.neutron [req-aa3a94f5-8a1f-4333-9fcf-b219d31769d4 req-2d367e74-123a-4124-8ac0-14099b4e4d38 service nova] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.391955] env[62914]: DEBUG nova.compute.utils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 767.394145] env[62914]: DEBUG nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 767.394974] env[62914]: DEBUG nova.network.neutron [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 767.411501] env[62914]: ERROR nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d34e9c7a-0dd0-4619-9d18-9fc611717c55, please check neutron logs for more information. [ 767.411501] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 767.411501] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.411501] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 767.411501] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.411501] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 767.411501] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.411501] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 767.411501] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.411501] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 767.411501] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.411501] env[62914]: ERROR nova.compute.manager raise self.value [ 767.411501] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.411501] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 767.411501] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.411501] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 767.412035] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.412035] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 767.412035] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d34e9c7a-0dd0-4619-9d18-9fc611717c55, please check neutron logs for more information. [ 767.412035] env[62914]: ERROR nova.compute.manager [ 767.412035] env[62914]: Traceback (most recent call last): [ 767.412035] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 767.412035] env[62914]: listener.cb(fileno) [ 767.412035] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.412035] env[62914]: result = function(*args, **kwargs) [ 767.412035] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 767.412035] env[62914]: return func(*args, **kwargs) [ 767.412035] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.412035] env[62914]: raise e [ 767.412035] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.412035] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 767.412035] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.412035] env[62914]: created_port_ids = self._update_ports_for_instance( [ 767.412035] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.412035] env[62914]: with excutils.save_and_reraise_exception(): [ 767.412035] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.412035] env[62914]: self.force_reraise() [ 767.412035] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.412035] env[62914]: raise self.value [ 767.412035] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.412035] env[62914]: updated_port = self._update_port( [ 767.412035] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.412035] env[62914]: _ensure_no_port_binding_failure(port) [ 767.412035] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.412035] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 767.412840] env[62914]: nova.exception.PortBindingFailed: Binding failed for port d34e9c7a-0dd0-4619-9d18-9fc611717c55, please check neutron logs for more information. [ 767.412840] env[62914]: Removing descriptor: 19 [ 767.412840] env[62914]: ERROR nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d34e9c7a-0dd0-4619-9d18-9fc611717c55, please check neutron logs for more information. [ 767.412840] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Traceback (most recent call last): [ 767.412840] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 767.412840] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] yield resources [ 767.412840] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 767.412840] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] self.driver.spawn(context, instance, image_meta, [ 767.412840] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 767.412840] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 767.412840] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 767.412840] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] vm_ref = self.build_virtual_machine(instance, [ 767.413693] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 767.413693] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] vif_infos = vmwarevif.get_vif_info(self._session, [ 767.413693] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 767.413693] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] for vif in network_info: [ 767.413693] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 767.413693] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] return self._sync_wrapper(fn, *args, **kwargs) [ 767.413693] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 767.413693] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] self.wait() [ 767.413693] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 767.413693] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] self[:] = self._gt.wait() [ 767.413693] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 767.413693] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] return self._exit_event.wait() [ 767.413693] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 767.414097] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] result = hub.switch() [ 767.414097] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 767.414097] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] return self.greenlet.switch() [ 767.414097] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.414097] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] result = function(*args, **kwargs) [ 767.414097] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 767.414097] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] return func(*args, **kwargs) [ 767.414097] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.414097] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] raise e [ 767.414097] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.414097] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] nwinfo = self.network_api.allocate_for_instance( [ 767.414097] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.414097] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] created_port_ids = self._update_ports_for_instance( [ 767.414473] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.414473] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] with excutils.save_and_reraise_exception(): [ 767.414473] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.414473] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] self.force_reraise() [ 767.414473] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.414473] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] raise self.value [ 767.414473] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.414473] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] updated_port = self._update_port( [ 767.414473] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.414473] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] _ensure_no_port_binding_failure(port) [ 767.414473] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.414473] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] raise exception.PortBindingFailed(port_id=port['id']) [ 767.415042] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] nova.exception.PortBindingFailed: Binding failed for port d34e9c7a-0dd0-4619-9d18-9fc611717c55, please check neutron logs for more information. [ 767.415042] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] [ 767.415042] env[62914]: INFO nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Terminating instance [ 767.416553] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Acquiring lock "refresh_cache-6201e8b7-ca3d-4eb0-8c45-67f54e66173c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.416553] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Acquired lock "refresh_cache-6201e8b7-ca3d-4eb0-8c45-67f54e66173c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.416771] env[62914]: DEBUG nova.network.neutron [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 767.458545] env[62914]: DEBUG nova.policy [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1d8b3f1914c24d3e9255ee1594120701', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '849460c330c14eabafe765b3d63a2675', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 767.543448] env[62914]: DEBUG oslo_concurrency.lockutils [req-aa3a94f5-8a1f-4333-9fcf-b219d31769d4 req-2d367e74-123a-4124-8ac0-14099b4e4d38 service nova] Releasing lock "refresh_cache-39489032-5430-471c-8d44-2c74682ed03b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.543448] env[62914]: DEBUG nova.compute.manager [req-aa3a94f5-8a1f-4333-9fcf-b219d31769d4 req-2d367e74-123a-4124-8ac0-14099b4e4d38 service nova] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Received event network-vif-deleted-6e9fb149-2aaa-4699-ac9d-371d56feba58 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 767.743061] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a3a32fe-8f2c-4003-90dc-0dfba7486eab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.743061] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577a2866-a445-4965-93e4-e8a55c357795 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.777203] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309c1207-3650-4486-a0d2-206bc822399c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.786877] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e10900-4e6e-455f-a826-87bba2f43e9f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.801024] env[62914]: DEBUG nova.compute.provider_tree [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.898202] env[62914]: DEBUG nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 767.953185] env[62914]: DEBUG nova.network.neutron [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.045691] env[62914]: DEBUG nova.network.neutron [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Successfully created port: 7e0d56a7-513e-4ab7-b310-b8ccf72cd976 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 768.095558] env[62914]: DEBUG nova.network.neutron [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.302941] env[62914]: DEBUG nova.scheduler.client.report [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.554940] env[62914]: DEBUG nova.compute.manager [req-c45be73f-154d-4bb4-a507-3e02bf81b3bf req-8659f747-bb3e-4e44-8518-0c53f3cd8750 service nova] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Received event network-changed-d34e9c7a-0dd0-4619-9d18-9fc611717c55 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.554940] env[62914]: DEBUG nova.compute.manager [req-c45be73f-154d-4bb4-a507-3e02bf81b3bf req-8659f747-bb3e-4e44-8518-0c53f3cd8750 service nova] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Refreshing instance network info cache due to event network-changed-d34e9c7a-0dd0-4619-9d18-9fc611717c55. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 768.554940] env[62914]: DEBUG oslo_concurrency.lockutils [req-c45be73f-154d-4bb4-a507-3e02bf81b3bf req-8659f747-bb3e-4e44-8518-0c53f3cd8750 service nova] Acquiring lock "refresh_cache-6201e8b7-ca3d-4eb0-8c45-67f54e66173c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.602071] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Releasing lock "refresh_cache-6201e8b7-ca3d-4eb0-8c45-67f54e66173c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.602071] env[62914]: DEBUG nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 768.602071] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 768.602071] env[62914]: DEBUG oslo_concurrency.lockutils [req-c45be73f-154d-4bb4-a507-3e02bf81b3bf req-8659f747-bb3e-4e44-8518-0c53f3cd8750 service nova] Acquired lock "refresh_cache-6201e8b7-ca3d-4eb0-8c45-67f54e66173c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.602071] env[62914]: DEBUG nova.network.neutron [req-c45be73f-154d-4bb4-a507-3e02bf81b3bf req-8659f747-bb3e-4e44-8518-0c53f3cd8750 service nova] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Refreshing network info cache for port d34e9c7a-0dd0-4619-9d18-9fc611717c55 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 768.602376] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b88c39b7-fbb7-4e3b-856a-db31d05e3a90 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.614245] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0390c037-ff63-4013-83fc-632c3a1d8c72 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.647032] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6201e8b7-ca3d-4eb0-8c45-67f54e66173c could not be found. [ 768.647032] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 768.647032] env[62914]: INFO nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 768.647032] env[62914]: DEBUG oslo.service.loopingcall [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.647032] env[62914]: DEBUG nova.compute.manager [-] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.647032] env[62914]: DEBUG nova.network.neutron [-] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 768.665464] env[62914]: DEBUG nova.network.neutron [-] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.813020] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.931s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.813020] env[62914]: ERROR nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f517bef4-915c-434b-95aa-baf973fd5446, please check neutron logs for more information. [ 768.813020] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Traceback (most recent call last): [ 768.813020] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 768.813020] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] self.driver.spawn(context, instance, image_meta, [ 768.813020] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 768.813020] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] self._vmops.spawn(context, instance, image_meta, injected_files, [ 768.813020] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 768.813020] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] vm_ref = self.build_virtual_machine(instance, [ 768.813520] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 768.813520] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] vif_infos = vmwarevif.get_vif_info(self._session, [ 768.813520] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 768.813520] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] for vif in network_info: [ 768.813520] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 768.813520] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] return self._sync_wrapper(fn, *args, **kwargs) [ 768.813520] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 768.813520] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] self.wait() [ 768.813520] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 768.813520] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] self[:] = self._gt.wait() [ 768.813520] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 768.813520] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] return self._exit_event.wait() [ 768.813520] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 768.813892] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] result = hub.switch() [ 768.813892] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 768.813892] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] return self.greenlet.switch() [ 768.813892] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 768.813892] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] result = function(*args, **kwargs) [ 768.813892] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 768.813892] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] return func(*args, **kwargs) [ 768.813892] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 768.813892] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] raise e [ 768.813892] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 768.813892] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] nwinfo = self.network_api.allocate_for_instance( [ 768.813892] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 768.813892] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] created_port_ids = self._update_ports_for_instance( [ 768.814279] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 768.814279] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] with excutils.save_and_reraise_exception(): [ 768.814279] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 768.814279] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] self.force_reraise() [ 768.814279] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 768.814279] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] raise self.value [ 768.814279] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 768.814279] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] updated_port = self._update_port( [ 768.814279] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 768.814279] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] _ensure_no_port_binding_failure(port) [ 768.814279] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 768.814279] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] raise exception.PortBindingFailed(port_id=port['id']) [ 768.814640] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] nova.exception.PortBindingFailed: Binding failed for port f517bef4-915c-434b-95aa-baf973fd5446, please check neutron logs for more information. [ 768.814640] env[62914]: ERROR nova.compute.manager [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] [ 768.814640] env[62914]: DEBUG nova.compute.utils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Binding failed for port f517bef4-915c-434b-95aa-baf973fd5446, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 768.814640] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.123s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.816484] env[62914]: DEBUG nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Build of instance 98c19874-7c71-4e20-aa29-2ddc53df7e81 was re-scheduled: Binding failed for port f517bef4-915c-434b-95aa-baf973fd5446, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 768.816942] env[62914]: DEBUG nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 768.817181] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "refresh_cache-98c19874-7c71-4e20-aa29-2ddc53df7e81" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.817322] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "refresh_cache-98c19874-7c71-4e20-aa29-2ddc53df7e81" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.817474] env[62914]: DEBUG nova.network.neutron [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 768.911020] env[62914]: DEBUG nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 768.936125] env[62914]: DEBUG nova.virt.hardware [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 768.936399] env[62914]: DEBUG nova.virt.hardware [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 768.936563] env[62914]: DEBUG nova.virt.hardware [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 768.936779] env[62914]: DEBUG nova.virt.hardware [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 768.936924] env[62914]: DEBUG nova.virt.hardware [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 768.937184] env[62914]: DEBUG nova.virt.hardware [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 768.937413] env[62914]: DEBUG nova.virt.hardware [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 768.937575] env[62914]: DEBUG nova.virt.hardware [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 768.937770] env[62914]: DEBUG nova.virt.hardware [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 768.937942] env[62914]: DEBUG nova.virt.hardware [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 768.938131] env[62914]: DEBUG nova.virt.hardware [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 768.939016] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f3b927-834a-4332-8e01-ac43186fcbf6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.947468] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aedbf82-bc10-4a22-8e7b-8b8083bfe4f1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.130928] env[62914]: DEBUG nova.network.neutron [req-c45be73f-154d-4bb4-a507-3e02bf81b3bf req-8659f747-bb3e-4e44-8518-0c53f3cd8750 service nova] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.169392] env[62914]: DEBUG nova.network.neutron [-] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.229167] env[62914]: DEBUG nova.network.neutron [req-c45be73f-154d-4bb4-a507-3e02bf81b3bf req-8659f747-bb3e-4e44-8518-0c53f3cd8750 service nova] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.270725] env[62914]: DEBUG nova.compute.manager [req-bf3a3f38-465a-4a7e-8034-005e56cac73e req-08292bcf-c6e3-4a8b-b138-96e9eb763395 service nova] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Received event network-changed-7e0d56a7-513e-4ab7-b310-b8ccf72cd976 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.270991] env[62914]: DEBUG nova.compute.manager [req-bf3a3f38-465a-4a7e-8034-005e56cac73e req-08292bcf-c6e3-4a8b-b138-96e9eb763395 service nova] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Refreshing instance network info cache due to event network-changed-7e0d56a7-513e-4ab7-b310-b8ccf72cd976. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 769.271466] env[62914]: DEBUG oslo_concurrency.lockutils [req-bf3a3f38-465a-4a7e-8034-005e56cac73e req-08292bcf-c6e3-4a8b-b138-96e9eb763395 service nova] Acquiring lock "refresh_cache-aa1e66d1-03fb-4340-83f1-eb79c29b6934" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.271674] env[62914]: DEBUG oslo_concurrency.lockutils [req-bf3a3f38-465a-4a7e-8034-005e56cac73e req-08292bcf-c6e3-4a8b-b138-96e9eb763395 service nova] Acquired lock "refresh_cache-aa1e66d1-03fb-4340-83f1-eb79c29b6934" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.271899] env[62914]: DEBUG nova.network.neutron [req-bf3a3f38-465a-4a7e-8034-005e56cac73e req-08292bcf-c6e3-4a8b-b138-96e9eb763395 service nova] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Refreshing network info cache for port 7e0d56a7-513e-4ab7-b310-b8ccf72cd976 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 769.340210] env[62914]: DEBUG nova.network.neutron [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.412250] env[62914]: DEBUG nova.network.neutron [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.540930] env[62914]: ERROR nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7e0d56a7-513e-4ab7-b310-b8ccf72cd976, please check neutron logs for more information. [ 769.540930] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 769.540930] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.540930] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 769.540930] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.540930] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 769.540930] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.540930] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 769.540930] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.540930] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 769.540930] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.540930] env[62914]: ERROR nova.compute.manager raise self.value [ 769.540930] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.540930] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 769.540930] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.540930] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 769.541380] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.541380] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 769.541380] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7e0d56a7-513e-4ab7-b310-b8ccf72cd976, please check neutron logs for more information. [ 769.541380] env[62914]: ERROR nova.compute.manager [ 769.541380] env[62914]: Traceback (most recent call last): [ 769.541380] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 769.541380] env[62914]: listener.cb(fileno) [ 769.541380] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.541380] env[62914]: result = function(*args, **kwargs) [ 769.541380] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 769.541380] env[62914]: return func(*args, **kwargs) [ 769.541380] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.541380] env[62914]: raise e [ 769.541380] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.541380] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 769.541380] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.541380] env[62914]: created_port_ids = self._update_ports_for_instance( [ 769.541380] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.541380] env[62914]: with excutils.save_and_reraise_exception(): [ 769.541380] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.541380] env[62914]: self.force_reraise() [ 769.541380] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.541380] env[62914]: raise self.value [ 769.541380] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.541380] env[62914]: updated_port = self._update_port( [ 769.541380] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.541380] env[62914]: _ensure_no_port_binding_failure(port) [ 769.541380] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.541380] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 769.542007] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 7e0d56a7-513e-4ab7-b310-b8ccf72cd976, please check neutron logs for more information. [ 769.542007] env[62914]: Removing descriptor: 15 [ 769.542078] env[62914]: ERROR nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7e0d56a7-513e-4ab7-b310-b8ccf72cd976, please check neutron logs for more information. [ 769.542078] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Traceback (most recent call last): [ 769.542078] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 769.542078] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] yield resources [ 769.542078] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.542078] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] self.driver.spawn(context, instance, image_meta, [ 769.542078] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 769.542078] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.542078] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.542078] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] vm_ref = self.build_virtual_machine(instance, [ 769.542078] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.542311] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.542311] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.542311] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] for vif in network_info: [ 769.542311] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.542311] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] return self._sync_wrapper(fn, *args, **kwargs) [ 769.542311] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.542311] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] self.wait() [ 769.542311] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.542311] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] self[:] = self._gt.wait() [ 769.542311] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.542311] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] return self._exit_event.wait() [ 769.542311] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 769.542311] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] result = hub.switch() [ 769.542591] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 769.542591] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] return self.greenlet.switch() [ 769.542591] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.542591] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] result = function(*args, **kwargs) [ 769.542591] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 769.542591] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] return func(*args, **kwargs) [ 769.542591] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.542591] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] raise e [ 769.542591] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.542591] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] nwinfo = self.network_api.allocate_for_instance( [ 769.542591] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.542591] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] created_port_ids = self._update_ports_for_instance( [ 769.542591] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.542843] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] with excutils.save_and_reraise_exception(): [ 769.542843] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.542843] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] self.force_reraise() [ 769.542843] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.542843] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] raise self.value [ 769.542843] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.542843] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] updated_port = self._update_port( [ 769.542843] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.542843] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] _ensure_no_port_binding_failure(port) [ 769.542843] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.542843] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] raise exception.PortBindingFailed(port_id=port['id']) [ 769.542843] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] nova.exception.PortBindingFailed: Binding failed for port 7e0d56a7-513e-4ab7-b310-b8ccf72cd976, please check neutron logs for more information. [ 769.542843] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] [ 769.543116] env[62914]: INFO nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Terminating instance [ 769.545922] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Acquiring lock "refresh_cache-aa1e66d1-03fb-4340-83f1-eb79c29b6934" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.673130] env[62914]: INFO nova.compute.manager [-] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Took 1.03 seconds to deallocate network for instance. [ 769.676310] env[62914]: DEBUG nova.compute.claims [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 769.676310] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.738641] env[62914]: DEBUG oslo_concurrency.lockutils [req-c45be73f-154d-4bb4-a507-3e02bf81b3bf req-8659f747-bb3e-4e44-8518-0c53f3cd8750 service nova] Releasing lock "refresh_cache-6201e8b7-ca3d-4eb0-8c45-67f54e66173c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.738641] env[62914]: DEBUG nova.compute.manager [req-c45be73f-154d-4bb4-a507-3e02bf81b3bf req-8659f747-bb3e-4e44-8518-0c53f3cd8750 service nova] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Received event network-vif-deleted-d34e9c7a-0dd0-4619-9d18-9fc611717c55 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.790832] env[62914]: DEBUG nova.network.neutron [req-bf3a3f38-465a-4a7e-8034-005e56cac73e req-08292bcf-c6e3-4a8b-b138-96e9eb763395 service nova] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.863649] env[62914]: WARNING nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 44339b45-5183-4882-aaec-4070adc3c3dd is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 769.915032] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "refresh_cache-98c19874-7c71-4e20-aa29-2ddc53df7e81" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.917577] env[62914]: DEBUG nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 769.917577] env[62914]: DEBUG nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 769.917577] env[62914]: DEBUG nova.network.neutron [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 769.930410] env[62914]: DEBUG nova.network.neutron [req-bf3a3f38-465a-4a7e-8034-005e56cac73e req-08292bcf-c6e3-4a8b-b138-96e9eb763395 service nova] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.944890] env[62914]: DEBUG nova.network.neutron [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.366753] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 98c19874-7c71-4e20-aa29-2ddc53df7e81 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.366955] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 364c5d87-f261-439a-920e-874ee0bcf83b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.367109] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 6201e8b7-ca3d-4eb0-8c45-67f54e66173c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.367237] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 39489032-5430-471c-8d44-2c74682ed03b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.367364] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance aa1e66d1-03fb-4340-83f1-eb79c29b6934 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.431681] env[62914]: DEBUG oslo_concurrency.lockutils [req-bf3a3f38-465a-4a7e-8034-005e56cac73e req-08292bcf-c6e3-4a8b-b138-96e9eb763395 service nova] Releasing lock "refresh_cache-aa1e66d1-03fb-4340-83f1-eb79c29b6934" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.432115] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Acquired lock "refresh_cache-aa1e66d1-03fb-4340-83f1-eb79c29b6934" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.432315] env[62914]: DEBUG nova.network.neutron [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 770.447766] env[62914]: DEBUG nova.network.neutron [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.876319] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 1d12cd15-22e9-4f96-8bd9-546196cb4138 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.951094] env[62914]: INFO nova.compute.manager [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 98c19874-7c71-4e20-aa29-2ddc53df7e81] Took 1.04 seconds to deallocate network for instance. [ 770.954812] env[62914]: DEBUG nova.network.neutron [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.069513] env[62914]: DEBUG nova.network.neutron [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.296540] env[62914]: DEBUG nova.compute.manager [req-b88475be-6ac0-46a9-89ae-9bf7d9e26d9a req-08d7dcf0-7841-473f-8f71-81571faf3a47 service nova] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Received event network-vif-deleted-7e0d56a7-513e-4ab7-b310-b8ccf72cd976 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.379807] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance dcf6b4e6-3d18-4afa-be0b-a18a58a13c08 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 771.486474] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Acquiring lock "bd3e82ce-62e7-4761-bcd7-d1f835234b97" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.486714] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Lock "bd3e82ce-62e7-4761-bcd7-d1f835234b97" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.574492] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Releasing lock "refresh_cache-aa1e66d1-03fb-4340-83f1-eb79c29b6934" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.574972] env[62914]: DEBUG nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 771.575181] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 771.575472] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5a8a1885-b540-4ff3-9708-360de6e83c60 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.584114] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a255b918-6ba6-4ead-9a62-6544ce2c6a0e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.604939] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aa1e66d1-03fb-4340-83f1-eb79c29b6934 could not be found. [ 771.605147] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 771.605327] env[62914]: INFO nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Took 0.03 seconds to destroy the instance on the hypervisor. [ 771.605561] env[62914]: DEBUG oslo.service.loopingcall [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.605790] env[62914]: DEBUG nova.compute.manager [-] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.605885] env[62914]: DEBUG nova.network.neutron [-] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 771.621920] env[62914]: DEBUG nova.network.neutron [-] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.884956] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 771.998021] env[62914]: INFO nova.scheduler.client.report [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleted allocations for instance 98c19874-7c71-4e20-aa29-2ddc53df7e81 [ 772.125128] env[62914]: DEBUG nova.network.neutron [-] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.388654] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance cd5a266b-617d-47c5-adb9-d161bd68f672 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.505588] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8d8adc3f-d185-4b7c-b79a-8679d1fbe31f tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "98c19874-7c71-4e20-aa29-2ddc53df7e81" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 165.273s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.628244] env[62914]: INFO nova.compute.manager [-] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Took 1.02 seconds to deallocate network for instance. [ 772.630661] env[62914]: DEBUG nova.compute.claims [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 772.630929] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.891835] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 6e38d30c-2fa3-468d-8ff9-e3966333c9c8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.008706] env[62914]: DEBUG nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 773.394687] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 8a342650-70f0-4d5b-8086-6450dcc8d95b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.533872] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.898816] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 87ef6b06-f699-4de3-8b89-854717074406 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 774.402817] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 9fab500c-344c-46cd-b060-c8f1aa0f1cba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 774.625952] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "2274ca27-8e2d-435e-8570-97eb4ae3cc58" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.626206] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "2274ca27-8e2d-435e-8570-97eb4ae3cc58" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.905886] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 775.409317] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 9fec3e97-4203-4a90-8a43-600d11d7e7a0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 775.912515] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 46d139a1-a4ae-435e-9d6f-cfc06d706128 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 776.415088] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance cb89c84f-414b-4dc7-9db5-12cbc30b52b4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 776.920068] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 9577394d-1c73-4ed1-ba86-e7c246e32719 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 777.423411] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 93d420a1-6d8f-4919-a42f-55aebab853ae has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 777.423411] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 777.423411] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 777.749172] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b589dcfe-e4c6-440c-a66b-ea3c8b93eaf1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.756471] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc68eae3-702a-44c0-a1e8-28ecafe77574 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.785424] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa32493-35cd-4ba3-9fc6-58b055c80c75 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.792157] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9bcfe6-b5ad-4a42-b2c2-c845df3af544 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.804735] env[62914]: DEBUG nova.compute.provider_tree [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.308467] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.813976] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 778.814270] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.002s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.814549] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.154s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.816084] env[62914]: INFO nova.compute.claims [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 778.818802] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 778.818949] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Cleaning up deleted instances {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 779.325618] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] There are 5 instances to clean {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 779.326018] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: e11079b9-5f15-41d6-8c8a-c08dc7f63b2f] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 779.831415] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: f3509973-013e-45ce-87f9-357e782f26d7] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 780.048832] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc2c18b-9588-41bf-a7d4-536de41864ef {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.056384] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6177559-443c-43f5-a436-47b77e662dab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.086201] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7b111f-083c-46a6-921a-9ec2bd242839 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.093239] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48445fb2-d200-482b-9ed6-67503dcd109e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.106022] env[62914]: DEBUG nova.compute.provider_tree [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.334565] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 34b0b81e-fc98-4019-92c3-fe2ce2fdfd67] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 780.608717] env[62914]: DEBUG nova.scheduler.client.report [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.837847] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: d5c90d13-41bc-4cca-abf3-5a5b13171a6f] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 781.113556] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.114054] env[62914]: DEBUG nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 781.116629] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.620s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.118102] env[62914]: INFO nova.compute.claims [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.341530] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 3786cd93-1a50-4997-8894-840afa1c8417] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 781.622313] env[62914]: DEBUG nova.compute.utils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 781.625567] env[62914]: DEBUG nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 781.625730] env[62914]: DEBUG nova.network.neutron [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 781.703283] env[62914]: DEBUG nova.policy [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '456f87454d3c484ea80e8ddf6b565cfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8e87bdcbeea4e9ea2b08630acc32ab1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 781.845289] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 781.845466] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Cleaning up deleted instances with incomplete migration {{(pid=62914) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 782.043134] env[62914]: DEBUG nova.network.neutron [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Successfully created port: 9bd49c11-a193-491f-ac82-dfb5b731f817 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 782.126634] env[62914]: DEBUG nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 782.347615] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.373315] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe5674d-9a11-40cc-bc94-29840303e4ce {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.381194] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceedecf2-605d-4786-b3ac-47defab890e7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.412152] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28011e85-7ff4-49e5-940b-2bb5887e2704 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.419473] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b3a063-ba7f-4840-be13-db3dd287d64b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.433319] env[62914]: DEBUG nova.compute.provider_tree [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.686172] env[62914]: DEBUG nova.compute.manager [req-ad4d494a-86fb-44c7-b27e-2ed14910f1c7 req-f552438b-15f4-40b3-b6b2-4d76ac6db40f service nova] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Received event network-changed-9bd49c11-a193-491f-ac82-dfb5b731f817 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.686402] env[62914]: DEBUG nova.compute.manager [req-ad4d494a-86fb-44c7-b27e-2ed14910f1c7 req-f552438b-15f4-40b3-b6b2-4d76ac6db40f service nova] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Refreshing instance network info cache due to event network-changed-9bd49c11-a193-491f-ac82-dfb5b731f817. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 782.686687] env[62914]: DEBUG oslo_concurrency.lockutils [req-ad4d494a-86fb-44c7-b27e-2ed14910f1c7 req-f552438b-15f4-40b3-b6b2-4d76ac6db40f service nova] Acquiring lock "refresh_cache-1d12cd15-22e9-4f96-8bd9-546196cb4138" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.686772] env[62914]: DEBUG oslo_concurrency.lockutils [req-ad4d494a-86fb-44c7-b27e-2ed14910f1c7 req-f552438b-15f4-40b3-b6b2-4d76ac6db40f service nova] Acquired lock "refresh_cache-1d12cd15-22e9-4f96-8bd9-546196cb4138" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.687617] env[62914]: DEBUG nova.network.neutron [req-ad4d494a-86fb-44c7-b27e-2ed14910f1c7 req-f552438b-15f4-40b3-b6b2-4d76ac6db40f service nova] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Refreshing network info cache for port 9bd49c11-a193-491f-ac82-dfb5b731f817 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 782.892969] env[62914]: ERROR nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9bd49c11-a193-491f-ac82-dfb5b731f817, please check neutron logs for more information. [ 782.892969] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 782.892969] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.892969] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 782.892969] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 782.892969] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 782.892969] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 782.892969] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 782.892969] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.892969] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 782.892969] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.892969] env[62914]: ERROR nova.compute.manager raise self.value [ 782.892969] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 782.892969] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 782.892969] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.892969] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 782.893366] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.893366] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 782.893366] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9bd49c11-a193-491f-ac82-dfb5b731f817, please check neutron logs for more information. [ 782.893366] env[62914]: ERROR nova.compute.manager [ 782.893366] env[62914]: Traceback (most recent call last): [ 782.893366] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 782.893366] env[62914]: listener.cb(fileno) [ 782.893366] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 782.893366] env[62914]: result = function(*args, **kwargs) [ 782.893366] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 782.893366] env[62914]: return func(*args, **kwargs) [ 782.893366] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 782.893366] env[62914]: raise e [ 782.893366] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.893366] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 782.893366] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 782.893366] env[62914]: created_port_ids = self._update_ports_for_instance( [ 782.893366] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 782.893366] env[62914]: with excutils.save_and_reraise_exception(): [ 782.893366] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.893366] env[62914]: self.force_reraise() [ 782.893366] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.893366] env[62914]: raise self.value [ 782.893366] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 782.893366] env[62914]: updated_port = self._update_port( [ 782.893366] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.893366] env[62914]: _ensure_no_port_binding_failure(port) [ 782.893366] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.893366] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 782.894035] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 9bd49c11-a193-491f-ac82-dfb5b731f817, please check neutron logs for more information. [ 782.894035] env[62914]: Removing descriptor: 19 [ 782.937015] env[62914]: DEBUG nova.scheduler.client.report [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.139443] env[62914]: DEBUG nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 783.164453] env[62914]: DEBUG nova.virt.hardware [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 783.164695] env[62914]: DEBUG nova.virt.hardware [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 783.164848] env[62914]: DEBUG nova.virt.hardware [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 783.165049] env[62914]: DEBUG nova.virt.hardware [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 783.165202] env[62914]: DEBUG nova.virt.hardware [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 783.165346] env[62914]: DEBUG nova.virt.hardware [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 783.165548] env[62914]: DEBUG nova.virt.hardware [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 783.165703] env[62914]: DEBUG nova.virt.hardware [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 783.165927] env[62914]: DEBUG nova.virt.hardware [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 783.166123] env[62914]: DEBUG nova.virt.hardware [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 783.166292] env[62914]: DEBUG nova.virt.hardware [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 783.167335] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d6cfac-7f18-4269-9f8f-dbf155bdfc73 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.175227] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a498c9c8-82fb-430d-a48f-42873e271415 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.188795] env[62914]: ERROR nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9bd49c11-a193-491f-ac82-dfb5b731f817, please check neutron logs for more information. [ 783.188795] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Traceback (most recent call last): [ 783.188795] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 783.188795] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] yield resources [ 783.188795] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 783.188795] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] self.driver.spawn(context, instance, image_meta, [ 783.188795] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 783.188795] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] self._vmops.spawn(context, instance, image_meta, injected_files, [ 783.188795] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 783.188795] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] vm_ref = self.build_virtual_machine(instance, [ 783.188795] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 783.189088] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] vif_infos = vmwarevif.get_vif_info(self._session, [ 783.189088] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 783.189088] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] for vif in network_info: [ 783.189088] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 783.189088] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] return self._sync_wrapper(fn, *args, **kwargs) [ 783.189088] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 783.189088] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] self.wait() [ 783.189088] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 783.189088] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] self[:] = self._gt.wait() [ 783.189088] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 783.189088] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] return self._exit_event.wait() [ 783.189088] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 783.189088] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] current.throw(*self._exc) [ 783.189355] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.189355] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] result = function(*args, **kwargs) [ 783.189355] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 783.189355] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] return func(*args, **kwargs) [ 783.189355] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 783.189355] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] raise e [ 783.189355] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 783.189355] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] nwinfo = self.network_api.allocate_for_instance( [ 783.189355] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.189355] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] created_port_ids = self._update_ports_for_instance( [ 783.189355] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.189355] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] with excutils.save_and_reraise_exception(): [ 783.189355] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.189631] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] self.force_reraise() [ 783.189631] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.189631] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] raise self.value [ 783.189631] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.189631] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] updated_port = self._update_port( [ 783.189631] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.189631] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] _ensure_no_port_binding_failure(port) [ 783.189631] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.189631] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] raise exception.PortBindingFailed(port_id=port['id']) [ 783.189631] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] nova.exception.PortBindingFailed: Binding failed for port 9bd49c11-a193-491f-ac82-dfb5b731f817, please check neutron logs for more information. [ 783.189631] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] [ 783.189631] env[62914]: INFO nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Terminating instance [ 783.191606] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Acquiring lock "refresh_cache-1d12cd15-22e9-4f96-8bd9-546196cb4138" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.211652] env[62914]: DEBUG nova.network.neutron [req-ad4d494a-86fb-44c7-b27e-2ed14910f1c7 req-f552438b-15f4-40b3-b6b2-4d76ac6db40f service nova] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.261032] env[62914]: DEBUG nova.network.neutron [req-ad4d494a-86fb-44c7-b27e-2ed14910f1c7 req-f552438b-15f4-40b3-b6b2-4d76ac6db40f service nova] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.440899] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.441435] env[62914]: DEBUG nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 783.444273] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.115s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.445780] env[62914]: INFO nova.compute.claims [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.763222] env[62914]: DEBUG oslo_concurrency.lockutils [req-ad4d494a-86fb-44c7-b27e-2ed14910f1c7 req-f552438b-15f4-40b3-b6b2-4d76ac6db40f service nova] Releasing lock "refresh_cache-1d12cd15-22e9-4f96-8bd9-546196cb4138" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.763648] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Acquired lock "refresh_cache-1d12cd15-22e9-4f96-8bd9-546196cb4138" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.763833] env[62914]: DEBUG nova.network.neutron [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.950735] env[62914]: DEBUG nova.compute.utils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 783.954069] env[62914]: DEBUG nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 783.954235] env[62914]: DEBUG nova.network.neutron [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 784.016297] env[62914]: DEBUG nova.policy [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c005e6cf21d9416db59047e5ebad0ef8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bd7fa76a52f4435912794f073d78fc8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 784.282498] env[62914]: DEBUG nova.network.neutron [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.329464] env[62914]: DEBUG nova.network.neutron [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.338176] env[62914]: DEBUG nova.network.neutron [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Successfully created port: 86111e45-f2b4-4aca-82fc-922d5fa3dd84 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 784.455093] env[62914]: DEBUG nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 784.715878] env[62914]: DEBUG nova.compute.manager [req-1a7e3073-1b34-493a-a848-eb3af32c4696 req-06e12763-7e61-4be6-bc50-30d36788ce4c service nova] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Received event network-vif-deleted-9bd49c11-a193-491f-ac82-dfb5b731f817 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.766626] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77740112-4794-49db-b23d-1e4253cba712 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.774334] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923a2d56-c113-47e0-be93-0e5b15c595b5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.803472] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d761670e-e7be-4d31-9b99-66f3a9a60e8e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.810708] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3bbeb48-e3fa-4e8d-a96b-e2fbd0e9fc01 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.823547] env[62914]: DEBUG nova.compute.provider_tree [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.831682] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Releasing lock "refresh_cache-1d12cd15-22e9-4f96-8bd9-546196cb4138" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.832068] env[62914]: DEBUG nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 784.832258] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 784.832495] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c1df1dbe-606c-4895-bc93-849a45776173 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.841928] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4dac0a9-9dae-4dc2-90d7-13d541a210f6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.863845] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1d12cd15-22e9-4f96-8bd9-546196cb4138 could not be found. [ 784.864064] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 784.864279] env[62914]: INFO nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Took 0.03 seconds to destroy the instance on the hypervisor. [ 784.864521] env[62914]: DEBUG oslo.service.loopingcall [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.864621] env[62914]: DEBUG nova.compute.manager [-] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.864713] env[62914]: DEBUG nova.network.neutron [-] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.880604] env[62914]: DEBUG nova.network.neutron [-] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.284935] env[62914]: ERROR nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 86111e45-f2b4-4aca-82fc-922d5fa3dd84, please check neutron logs for more information. [ 785.284935] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 785.284935] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 785.284935] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 785.284935] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 785.284935] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 785.284935] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 785.284935] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 785.284935] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 785.284935] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 785.284935] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 785.284935] env[62914]: ERROR nova.compute.manager raise self.value [ 785.284935] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 785.284935] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 785.284935] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 785.284935] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 785.285402] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 785.285402] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 785.285402] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 86111e45-f2b4-4aca-82fc-922d5fa3dd84, please check neutron logs for more information. [ 785.285402] env[62914]: ERROR nova.compute.manager [ 785.285402] env[62914]: Traceback (most recent call last): [ 785.285402] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 785.285402] env[62914]: listener.cb(fileno) [ 785.285402] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 785.285402] env[62914]: result = function(*args, **kwargs) [ 785.285402] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 785.285402] env[62914]: return func(*args, **kwargs) [ 785.285402] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 785.285402] env[62914]: raise e [ 785.285402] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 785.285402] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 785.285402] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 785.285402] env[62914]: created_port_ids = self._update_ports_for_instance( [ 785.285402] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 785.285402] env[62914]: with excutils.save_and_reraise_exception(): [ 785.285402] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 785.285402] env[62914]: self.force_reraise() [ 785.285402] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 785.285402] env[62914]: raise self.value [ 785.285402] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 785.285402] env[62914]: updated_port = self._update_port( [ 785.285402] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 785.285402] env[62914]: _ensure_no_port_binding_failure(port) [ 785.285402] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 785.285402] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 785.286065] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 86111e45-f2b4-4aca-82fc-922d5fa3dd84, please check neutron logs for more information. [ 785.286065] env[62914]: Removing descriptor: 19 [ 785.329036] env[62914]: DEBUG nova.scheduler.client.report [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.383282] env[62914]: DEBUG nova.network.neutron [-] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.467505] env[62914]: DEBUG nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 785.491131] env[62914]: DEBUG nova.virt.hardware [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 785.491368] env[62914]: DEBUG nova.virt.hardware [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 785.491517] env[62914]: DEBUG nova.virt.hardware [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 785.491689] env[62914]: DEBUG nova.virt.hardware [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 785.491826] env[62914]: DEBUG nova.virt.hardware [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 785.491964] env[62914]: DEBUG nova.virt.hardware [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 785.492181] env[62914]: DEBUG nova.virt.hardware [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 785.492333] env[62914]: DEBUG nova.virt.hardware [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 785.492491] env[62914]: DEBUG nova.virt.hardware [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 785.492645] env[62914]: DEBUG nova.virt.hardware [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 785.492805] env[62914]: DEBUG nova.virt.hardware [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 785.493720] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d188af3c-301b-4637-b7fb-3699b833bf0e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.501805] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea9122f9-3f57-46ea-a874-baf571310f32 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.514595] env[62914]: ERROR nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 86111e45-f2b4-4aca-82fc-922d5fa3dd84, please check neutron logs for more information. [ 785.514595] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Traceback (most recent call last): [ 785.514595] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 785.514595] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] yield resources [ 785.514595] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 785.514595] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] self.driver.spawn(context, instance, image_meta, [ 785.514595] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 785.514595] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] self._vmops.spawn(context, instance, image_meta, injected_files, [ 785.514595] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 785.514595] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] vm_ref = self.build_virtual_machine(instance, [ 785.514595] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 785.514912] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] vif_infos = vmwarevif.get_vif_info(self._session, [ 785.514912] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 785.514912] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] for vif in network_info: [ 785.514912] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 785.514912] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] return self._sync_wrapper(fn, *args, **kwargs) [ 785.514912] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 785.514912] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] self.wait() [ 785.514912] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 785.514912] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] self[:] = self._gt.wait() [ 785.514912] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 785.514912] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] return self._exit_event.wait() [ 785.514912] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 785.514912] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] current.throw(*self._exc) [ 785.515257] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 785.515257] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] result = function(*args, **kwargs) [ 785.515257] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 785.515257] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] return func(*args, **kwargs) [ 785.515257] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 785.515257] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] raise e [ 785.515257] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 785.515257] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] nwinfo = self.network_api.allocate_for_instance( [ 785.515257] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 785.515257] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] created_port_ids = self._update_ports_for_instance( [ 785.515257] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 785.515257] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] with excutils.save_and_reraise_exception(): [ 785.515257] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 785.515574] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] self.force_reraise() [ 785.515574] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 785.515574] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] raise self.value [ 785.515574] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 785.515574] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] updated_port = self._update_port( [ 785.515574] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 785.515574] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] _ensure_no_port_binding_failure(port) [ 785.515574] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 785.515574] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] raise exception.PortBindingFailed(port_id=port['id']) [ 785.515574] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] nova.exception.PortBindingFailed: Binding failed for port 86111e45-f2b4-4aca-82fc-922d5fa3dd84, please check neutron logs for more information. [ 785.515574] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] [ 785.515574] env[62914]: INFO nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Terminating instance [ 785.516723] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Acquiring lock "refresh_cache-dcf6b4e6-3d18-4afa-be0b-a18a58a13c08" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.516878] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Acquired lock "refresh_cache-dcf6b4e6-3d18-4afa-be0b-a18a58a13c08" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.517062] env[62914]: DEBUG nova.network.neutron [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 785.833609] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.834008] env[62914]: DEBUG nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.836855] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.638s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.838372] env[62914]: INFO nova.compute.claims [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 785.885431] env[62914]: INFO nova.compute.manager [-] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Took 1.02 seconds to deallocate network for instance. [ 785.887582] env[62914]: DEBUG nova.compute.claims [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 785.887752] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.034345] env[62914]: DEBUG nova.network.neutron [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.122440] env[62914]: DEBUG nova.network.neutron [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.343208] env[62914]: DEBUG nova.compute.utils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.346857] env[62914]: DEBUG nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 786.347040] env[62914]: DEBUG nova.network.neutron [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 786.395976] env[62914]: DEBUG nova.policy [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1c239f8c93f04578b2460466a9aea7d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c1e89c454a74089a470b8d322e1aea7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 786.625459] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Releasing lock "refresh_cache-dcf6b4e6-3d18-4afa-be0b-a18a58a13c08" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.625780] env[62914]: DEBUG nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 786.626034] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 786.626321] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-940913b0-2d42-4fd0-8aff-95334849c5a2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.637988] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a06fc0d3-06b7-41f8-be67-825431bb2401 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.653796] env[62914]: DEBUG nova.network.neutron [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Successfully created port: bb5cc316-1f83-4dad-8218-d3f8bdacb51f {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.668999] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dcf6b4e6-3d18-4afa-be0b-a18a58a13c08 could not be found. [ 786.669315] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 786.669502] env[62914]: INFO nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Took 0.04 seconds to destroy the instance on the hypervisor. [ 786.669738] env[62914]: DEBUG oslo.service.loopingcall [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.669952] env[62914]: DEBUG nova.compute.manager [-] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 786.670059] env[62914]: DEBUG nova.network.neutron [-] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 786.685991] env[62914]: DEBUG nova.network.neutron [-] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.780441] env[62914]: DEBUG nova.compute.manager [req-3371eabf-2afc-42b0-b796-a3e4b78ce36f req-28e0019a-012d-45f3-a928-80ee706b7bd9 service nova] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Received event network-changed-86111e45-f2b4-4aca-82fc-922d5fa3dd84 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.781018] env[62914]: DEBUG nova.compute.manager [req-3371eabf-2afc-42b0-b796-a3e4b78ce36f req-28e0019a-012d-45f3-a928-80ee706b7bd9 service nova] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Refreshing instance network info cache due to event network-changed-86111e45-f2b4-4aca-82fc-922d5fa3dd84. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 786.781018] env[62914]: DEBUG oslo_concurrency.lockutils [req-3371eabf-2afc-42b0-b796-a3e4b78ce36f req-28e0019a-012d-45f3-a928-80ee706b7bd9 service nova] Acquiring lock "refresh_cache-dcf6b4e6-3d18-4afa-be0b-a18a58a13c08" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.781018] env[62914]: DEBUG oslo_concurrency.lockutils [req-3371eabf-2afc-42b0-b796-a3e4b78ce36f req-28e0019a-012d-45f3-a928-80ee706b7bd9 service nova] Acquired lock "refresh_cache-dcf6b4e6-3d18-4afa-be0b-a18a58a13c08" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.781254] env[62914]: DEBUG nova.network.neutron [req-3371eabf-2afc-42b0-b796-a3e4b78ce36f req-28e0019a-012d-45f3-a928-80ee706b7bd9 service nova] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Refreshing network info cache for port 86111e45-f2b4-4aca-82fc-922d5fa3dd84 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 786.847966] env[62914]: DEBUG nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 787.099901] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-878393aa-f1c7-4f4f-bd5d-817d35a70792 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.107632] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d7ba91-8c75-4d81-80a7-e658f6416b5e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.139401] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668339a0-4ff2-4408-bec8-c09bc213bc28 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.147248] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41abd15-3c3f-4efa-bd20-9c5c3da96408 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.161516] env[62914]: DEBUG nova.compute.provider_tree [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.188805] env[62914]: DEBUG nova.network.neutron [-] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.302698] env[62914]: DEBUG nova.network.neutron [req-3371eabf-2afc-42b0-b796-a3e4b78ce36f req-28e0019a-012d-45f3-a928-80ee706b7bd9 service nova] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.356201] env[62914]: INFO nova.virt.block_device [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Booting with volume 298c93e0-8514-4b64-a23a-4ce5e4594abe at /dev/sda [ 787.380096] env[62914]: DEBUG nova.network.neutron [req-3371eabf-2afc-42b0-b796-a3e4b78ce36f req-28e0019a-012d-45f3-a928-80ee706b7bd9 service nova] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.406148] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e085139-fae3-4bca-9080-c4aed2a67486 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.210602] env[62914]: ERROR nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bb5cc316-1f83-4dad-8218-d3f8bdacb51f, please check neutron logs for more information. [ 788.210602] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 788.210602] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 788.210602] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 788.210602] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 788.210602] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 788.210602] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 788.210602] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 788.210602] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 788.210602] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 788.210602] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 788.210602] env[62914]: ERROR nova.compute.manager raise self.value [ 788.210602] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 788.210602] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 788.210602] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 788.210602] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 788.211740] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 788.211740] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 788.211740] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bb5cc316-1f83-4dad-8218-d3f8bdacb51f, please check neutron logs for more information. [ 788.211740] env[62914]: ERROR nova.compute.manager [ 788.211740] env[62914]: Traceback (most recent call last): [ 788.211740] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 788.211740] env[62914]: listener.cb(fileno) [ 788.211740] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 788.211740] env[62914]: result = function(*args, **kwargs) [ 788.211740] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 788.211740] env[62914]: return func(*args, **kwargs) [ 788.211740] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 788.211740] env[62914]: raise e [ 788.211740] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 788.211740] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 788.211740] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 788.211740] env[62914]: created_port_ids = self._update_ports_for_instance( [ 788.211740] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 788.211740] env[62914]: with excutils.save_and_reraise_exception(): [ 788.211740] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 788.211740] env[62914]: self.force_reraise() [ 788.211740] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 788.211740] env[62914]: raise self.value [ 788.211740] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 788.211740] env[62914]: updated_port = self._update_port( [ 788.211740] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 788.211740] env[62914]: _ensure_no_port_binding_failure(port) [ 788.211740] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 788.211740] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 788.212658] env[62914]: nova.exception.PortBindingFailed: Binding failed for port bb5cc316-1f83-4dad-8218-d3f8bdacb51f, please check neutron logs for more information. [ 788.212658] env[62914]: Removing descriptor: 19 [ 788.212658] env[62914]: DEBUG nova.scheduler.client.report [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.215031] env[62914]: INFO nova.compute.manager [-] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Took 1.54 seconds to deallocate network for instance. [ 788.215774] env[62914]: DEBUG oslo_concurrency.lockutils [req-3371eabf-2afc-42b0-b796-a3e4b78ce36f req-28e0019a-012d-45f3-a928-80ee706b7bd9 service nova] Releasing lock "refresh_cache-dcf6b4e6-3d18-4afa-be0b-a18a58a13c08" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.216011] env[62914]: DEBUG nova.compute.manager [req-3371eabf-2afc-42b0-b796-a3e4b78ce36f req-28e0019a-012d-45f3-a928-80ee706b7bd9 service nova] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Received event network-vif-deleted-86111e45-f2b4-4aca-82fc-922d5fa3dd84 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.218846] env[62914]: DEBUG nova.compute.claims [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 788.218846] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.227161] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe464f6-9b0f-4822-a3ff-c5d7f2e28f6c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.246998] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ae94f91-bfbc-47d1-aff8-49a8d47ff8f9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.254554] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a571da3b-fae0-4b29-972c-1f78682d50ee {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.275438] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4cc44e-98df-41e6-8dbb-c7e98bdd2191 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.281724] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffbb6a7b-a002-4898-92ff-ffc7373c3e7b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.294580] env[62914]: DEBUG nova.virt.block_device [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Updating existing volume attachment record: 28f8b72b-86ca-4f3c-b309-6de4c71a2672 {{(pid=62914) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 788.715815] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.879s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.716459] env[62914]: DEBUG nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 788.719068] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 28.128s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.719766] env[62914]: DEBUG nova.objects.instance [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62914) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 788.806793] env[62914]: DEBUG nova.compute.manager [req-4cd95458-003c-4f9f-a606-8671a02fd4b8 req-458a52bd-99ce-4fe1-936f-c1227600a159 service nova] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Received event network-changed-bb5cc316-1f83-4dad-8218-d3f8bdacb51f {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.806876] env[62914]: DEBUG nova.compute.manager [req-4cd95458-003c-4f9f-a606-8671a02fd4b8 req-458a52bd-99ce-4fe1-936f-c1227600a159 service nova] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Refreshing instance network info cache due to event network-changed-bb5cc316-1f83-4dad-8218-d3f8bdacb51f. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 788.807408] env[62914]: DEBUG oslo_concurrency.lockutils [req-4cd95458-003c-4f9f-a606-8671a02fd4b8 req-458a52bd-99ce-4fe1-936f-c1227600a159 service nova] Acquiring lock "refresh_cache-cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.807566] env[62914]: DEBUG oslo_concurrency.lockutils [req-4cd95458-003c-4f9f-a606-8671a02fd4b8 req-458a52bd-99ce-4fe1-936f-c1227600a159 service nova] Acquired lock "refresh_cache-cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.807722] env[62914]: DEBUG nova.network.neutron [req-4cd95458-003c-4f9f-a606-8671a02fd4b8 req-458a52bd-99ce-4fe1-936f-c1227600a159 service nova] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Refreshing network info cache for port bb5cc316-1f83-4dad-8218-d3f8bdacb51f {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 789.226371] env[62914]: DEBUG nova.compute.utils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 789.227742] env[62914]: DEBUG nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 789.227907] env[62914]: DEBUG nova.network.neutron [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 789.284110] env[62914]: DEBUG nova.policy [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a26360a953ef4c4da7cf614c2c15c766', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4724e78ceebc424db58c1172bc2d1828', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 789.323654] env[62914]: DEBUG nova.network.neutron [req-4cd95458-003c-4f9f-a606-8671a02fd4b8 req-458a52bd-99ce-4fe1-936f-c1227600a159 service nova] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.405689] env[62914]: DEBUG nova.network.neutron [req-4cd95458-003c-4f9f-a606-8671a02fd4b8 req-458a52bd-99ce-4fe1-936f-c1227600a159 service nova] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.549389] env[62914]: DEBUG nova.network.neutron [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Successfully created port: 21368432-56cd-4055-9388-b3ab78e2bcb3 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.729636] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5fad663b-e118-4171-8611-d020895653c0 tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.730727] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.753s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.732142] env[62914]: INFO nova.compute.claims [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 789.734916] env[62914]: DEBUG nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 789.908255] env[62914]: DEBUG oslo_concurrency.lockutils [req-4cd95458-003c-4f9f-a606-8671a02fd4b8 req-458a52bd-99ce-4fe1-936f-c1227600a159 service nova] Releasing lock "refresh_cache-cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.908448] env[62914]: DEBUG nova.compute.manager [req-4cd95458-003c-4f9f-a606-8671a02fd4b8 req-458a52bd-99ce-4fe1-936f-c1227600a159 service nova] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Received event network-vif-deleted-bb5cc316-1f83-4dad-8218-d3f8bdacb51f {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.381405] env[62914]: DEBUG nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 790.381940] env[62914]: DEBUG nova.virt.hardware [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.382190] env[62914]: DEBUG nova.virt.hardware [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.382519] env[62914]: DEBUG nova.virt.hardware [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.382738] env[62914]: DEBUG nova.virt.hardware [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.382889] env[62914]: DEBUG nova.virt.hardware [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.383055] env[62914]: DEBUG nova.virt.hardware [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.383263] env[62914]: DEBUG nova.virt.hardware [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.383420] env[62914]: DEBUG nova.virt.hardware [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.383583] env[62914]: DEBUG nova.virt.hardware [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.383739] env[62914]: DEBUG nova.virt.hardware [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.383903] env[62914]: DEBUG nova.virt.hardware [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.384766] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de818b3b-befc-423d-9387-a2e36a393cf6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.393149] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99122915-a61a-440f-98a0-36be8e6ffcd1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.412210] env[62914]: ERROR nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bb5cc316-1f83-4dad-8218-d3f8bdacb51f, please check neutron logs for more information. [ 790.412210] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Traceback (most recent call last): [ 790.412210] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 790.412210] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] yield resources [ 790.412210] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 790.412210] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] self.driver.spawn(context, instance, image_meta, [ 790.412210] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 790.412210] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] self._vmops.spawn(context, instance, image_meta, injected_files, [ 790.412210] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 790.412210] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] vm_ref = self.build_virtual_machine(instance, [ 790.412210] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 790.412499] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] vif_infos = vmwarevif.get_vif_info(self._session, [ 790.412499] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 790.412499] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] for vif in network_info: [ 790.412499] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 790.412499] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] return self._sync_wrapper(fn, *args, **kwargs) [ 790.412499] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 790.412499] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] self.wait() [ 790.412499] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 790.412499] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] self[:] = self._gt.wait() [ 790.412499] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 790.412499] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] return self._exit_event.wait() [ 790.412499] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 790.412499] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] current.throw(*self._exc) [ 790.412834] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.412834] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] result = function(*args, **kwargs) [ 790.412834] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 790.412834] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] return func(*args, **kwargs) [ 790.412834] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 790.412834] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] raise e [ 790.412834] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.412834] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] nwinfo = self.network_api.allocate_for_instance( [ 790.412834] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.412834] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] created_port_ids = self._update_ports_for_instance( [ 790.412834] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.412834] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] with excutils.save_and_reraise_exception(): [ 790.412834] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.413126] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] self.force_reraise() [ 790.413126] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.413126] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] raise self.value [ 790.413126] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.413126] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] updated_port = self._update_port( [ 790.413126] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.413126] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] _ensure_no_port_binding_failure(port) [ 790.413126] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.413126] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] raise exception.PortBindingFailed(port_id=port['id']) [ 790.413126] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] nova.exception.PortBindingFailed: Binding failed for port bb5cc316-1f83-4dad-8218-d3f8bdacb51f, please check neutron logs for more information. [ 790.413126] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] [ 790.413126] env[62914]: INFO nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Terminating instance [ 790.414656] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Acquiring lock "refresh_cache-cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.414814] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Acquired lock "refresh_cache-cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.414977] env[62914]: DEBUG nova.network.neutron [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 790.563601] env[62914]: ERROR nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 21368432-56cd-4055-9388-b3ab78e2bcb3, please check neutron logs for more information. [ 790.563601] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 790.563601] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.563601] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 790.563601] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.563601] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 790.563601] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.563601] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 790.563601] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.563601] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 790.563601] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.563601] env[62914]: ERROR nova.compute.manager raise self.value [ 790.563601] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.563601] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 790.563601] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.563601] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 790.564137] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.564137] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 790.564137] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 21368432-56cd-4055-9388-b3ab78e2bcb3, please check neutron logs for more information. [ 790.564137] env[62914]: ERROR nova.compute.manager [ 790.564137] env[62914]: Traceback (most recent call last): [ 790.564137] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 790.564137] env[62914]: listener.cb(fileno) [ 790.564137] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.564137] env[62914]: result = function(*args, **kwargs) [ 790.564137] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 790.564137] env[62914]: return func(*args, **kwargs) [ 790.564137] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 790.564137] env[62914]: raise e [ 790.564137] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.564137] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 790.564137] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.564137] env[62914]: created_port_ids = self._update_ports_for_instance( [ 790.564137] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.564137] env[62914]: with excutils.save_and_reraise_exception(): [ 790.564137] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.564137] env[62914]: self.force_reraise() [ 790.564137] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.564137] env[62914]: raise self.value [ 790.564137] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.564137] env[62914]: updated_port = self._update_port( [ 790.564137] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.564137] env[62914]: _ensure_no_port_binding_failure(port) [ 790.564137] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.564137] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 790.564910] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 21368432-56cd-4055-9388-b3ab78e2bcb3, please check neutron logs for more information. [ 790.564910] env[62914]: Removing descriptor: 19 [ 790.749058] env[62914]: DEBUG nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 790.773710] env[62914]: DEBUG nova.virt.hardware [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.773961] env[62914]: DEBUG nova.virt.hardware [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.774130] env[62914]: DEBUG nova.virt.hardware [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.774312] env[62914]: DEBUG nova.virt.hardware [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.774448] env[62914]: DEBUG nova.virt.hardware [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.774588] env[62914]: DEBUG nova.virt.hardware [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.774786] env[62914]: DEBUG nova.virt.hardware [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.774938] env[62914]: DEBUG nova.virt.hardware [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.775128] env[62914]: DEBUG nova.virt.hardware [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.775280] env[62914]: DEBUG nova.virt.hardware [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.775443] env[62914]: DEBUG nova.virt.hardware [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.776300] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a4ae86-364f-436f-8d5a-6aa885d6cf25 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.785772] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e1fda19-57c0-4b87-82d7-fb4d67b53026 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.801887] env[62914]: ERROR nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 21368432-56cd-4055-9388-b3ab78e2bcb3, please check neutron logs for more information. [ 790.801887] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Traceback (most recent call last): [ 790.801887] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 790.801887] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] yield resources [ 790.801887] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 790.801887] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] self.driver.spawn(context, instance, image_meta, [ 790.801887] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 790.801887] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] self._vmops.spawn(context, instance, image_meta, injected_files, [ 790.801887] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 790.801887] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] vm_ref = self.build_virtual_machine(instance, [ 790.801887] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 790.802171] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] vif_infos = vmwarevif.get_vif_info(self._session, [ 790.802171] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 790.802171] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] for vif in network_info: [ 790.802171] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 790.802171] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] return self._sync_wrapper(fn, *args, **kwargs) [ 790.802171] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 790.802171] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] self.wait() [ 790.802171] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 790.802171] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] self[:] = self._gt.wait() [ 790.802171] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 790.802171] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] return self._exit_event.wait() [ 790.802171] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 790.802171] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] current.throw(*self._exc) [ 790.802447] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.802447] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] result = function(*args, **kwargs) [ 790.802447] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 790.802447] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] return func(*args, **kwargs) [ 790.802447] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 790.802447] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] raise e [ 790.802447] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.802447] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] nwinfo = self.network_api.allocate_for_instance( [ 790.802447] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.802447] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] created_port_ids = self._update_ports_for_instance( [ 790.802447] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.802447] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] with excutils.save_and_reraise_exception(): [ 790.802447] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.802779] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] self.force_reraise() [ 790.802779] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.802779] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] raise self.value [ 790.802779] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.802779] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] updated_port = self._update_port( [ 790.802779] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.802779] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] _ensure_no_port_binding_failure(port) [ 790.802779] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.802779] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] raise exception.PortBindingFailed(port_id=port['id']) [ 790.802779] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] nova.exception.PortBindingFailed: Binding failed for port 21368432-56cd-4055-9388-b3ab78e2bcb3, please check neutron logs for more information. [ 790.802779] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] [ 790.802779] env[62914]: INFO nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Terminating instance [ 790.805503] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Acquiring lock "refresh_cache-cd5a266b-617d-47c5-adb9-d161bd68f672" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.805666] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Acquired lock "refresh_cache-cd5a266b-617d-47c5-adb9-d161bd68f672" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.805865] env[62914]: DEBUG nova.network.neutron [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 790.832334] env[62914]: DEBUG nova.compute.manager [req-b0a52d2e-b623-41aa-a074-6f9a1a86d913 req-d33bf36b-a966-4fd2-a41f-13766638b5c9 service nova] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Received event network-changed-21368432-56cd-4055-9388-b3ab78e2bcb3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.832523] env[62914]: DEBUG nova.compute.manager [req-b0a52d2e-b623-41aa-a074-6f9a1a86d913 req-d33bf36b-a966-4fd2-a41f-13766638b5c9 service nova] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Refreshing instance network info cache due to event network-changed-21368432-56cd-4055-9388-b3ab78e2bcb3. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 790.832709] env[62914]: DEBUG oslo_concurrency.lockutils [req-b0a52d2e-b623-41aa-a074-6f9a1a86d913 req-d33bf36b-a966-4fd2-a41f-13766638b5c9 service nova] Acquiring lock "refresh_cache-cd5a266b-617d-47c5-adb9-d161bd68f672" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.932748] env[62914]: DEBUG nova.network.neutron [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.024339] env[62914]: DEBUG nova.network.neutron [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.058684] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3094af52-63e4-4b73-8488-82e57349108a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.067292] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75d733c7-cdfd-4752-83c6-769b684b17f9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.095827] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77a8279b-dc53-4899-a541-7e508dcf0ad8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.102818] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f69132d-7124-4824-a984-581ce7133246 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.115983] env[62914]: DEBUG nova.compute.provider_tree [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.321628] env[62914]: DEBUG nova.network.neutron [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.391727] env[62914]: DEBUG nova.network.neutron [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.527298] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Releasing lock "refresh_cache-cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.528197] env[62914]: DEBUG nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 791.529196] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4b40472e-4b56-4a6b-b6f0-dbd818b124f3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.538375] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef716b19-0c98-4cf4-8bd2-8ce27e7f7aec {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.558619] env[62914]: WARNING nova.virt.vmwareapi.driver [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206 could not be found. [ 791.558956] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 791.559356] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17323892-d691-4eaf-802c-75a7767d49b7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.566362] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb57cf8-4cf0-4536-ac74-facd2bf0bc63 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.587130] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206 could not be found. [ 791.587495] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 791.587807] env[62914]: INFO nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Took 0.06 seconds to destroy the instance on the hypervisor. [ 791.588384] env[62914]: DEBUG oslo.service.loopingcall [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.589051] env[62914]: DEBUG nova.compute.manager [-] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 791.589051] env[62914]: DEBUG nova.network.neutron [-] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 791.603761] env[62914]: DEBUG nova.network.neutron [-] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.619251] env[62914]: DEBUG nova.scheduler.client.report [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.896376] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Releasing lock "refresh_cache-cd5a266b-617d-47c5-adb9-d161bd68f672" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.896972] env[62914]: DEBUG nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 791.897081] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 791.897717] env[62914]: DEBUG oslo_concurrency.lockutils [req-b0a52d2e-b623-41aa-a074-6f9a1a86d913 req-d33bf36b-a966-4fd2-a41f-13766638b5c9 service nova] Acquired lock "refresh_cache-cd5a266b-617d-47c5-adb9-d161bd68f672" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.897896] env[62914]: DEBUG nova.network.neutron [req-b0a52d2e-b623-41aa-a074-6f9a1a86d913 req-d33bf36b-a966-4fd2-a41f-13766638b5c9 service nova] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Refreshing network info cache for port 21368432-56cd-4055-9388-b3ab78e2bcb3 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 791.898927] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f92273de-bb4f-40ac-843b-be312dded963 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.909021] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0723869-4a1d-4070-af4a-a1352592a621 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.929635] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cd5a266b-617d-47c5-adb9-d161bd68f672 could not be found. [ 791.929833] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 791.930014] env[62914]: INFO nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Took 0.03 seconds to destroy the instance on the hypervisor. [ 791.930250] env[62914]: DEBUG oslo.service.loopingcall [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.930461] env[62914]: DEBUG nova.compute.manager [-] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 791.930560] env[62914]: DEBUG nova.network.neutron [-] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 791.944773] env[62914]: DEBUG nova.network.neutron [-] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.106291] env[62914]: DEBUG nova.network.neutron [-] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.124638] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.125182] env[62914]: DEBUG nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 792.127609] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.389s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.127815] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.129788] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.667s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.149505] env[62914]: INFO nova.scheduler.client.report [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Deleted allocations for instance 44339b45-5183-4882-aaec-4070adc3c3dd [ 792.416891] env[62914]: DEBUG nova.network.neutron [req-b0a52d2e-b623-41aa-a074-6f9a1a86d913 req-d33bf36b-a966-4fd2-a41f-13766638b5c9 service nova] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.446987] env[62914]: DEBUG nova.network.neutron [-] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.485866] env[62914]: DEBUG nova.network.neutron [req-b0a52d2e-b623-41aa-a074-6f9a1a86d913 req-d33bf36b-a966-4fd2-a41f-13766638b5c9 service nova] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.609441] env[62914]: INFO nova.compute.manager [-] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Took 1.02 seconds to deallocate network for instance. [ 792.634729] env[62914]: DEBUG nova.compute.utils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 792.639011] env[62914]: DEBUG nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 792.639178] env[62914]: DEBUG nova.network.neutron [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 792.657041] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b4cf0ee-05f2-4fa5-9f0b-c27567f884ad tempest-ServerShowV257Test-1498331500 tempest-ServerShowV257Test-1498331500-project-member] Lock "44339b45-5183-4882-aaec-4070adc3c3dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.696s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.712438] env[62914]: DEBUG nova.policy [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce06463e715b4975bc43b14be64ef9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '619be7b3bf1445b68ba7adefc98f8782', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 792.916739] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239a56cc-d252-4616-b2d0-31871689d7c3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.924242] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20ad403-7429-43a0-b25e-a20abf836035 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.953698] env[62914]: INFO nova.compute.manager [-] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Took 1.02 seconds to deallocate network for instance. [ 792.956231] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc5ec7b-0c2f-45d7-a2ee-16d3f1e1c18b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.959119] env[62914]: DEBUG nova.compute.claims [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 792.959302] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.964433] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1a03fd-a2b2-482f-80b7-a7f10cb42935 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.977525] env[62914]: DEBUG nova.compute.provider_tree [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.988634] env[62914]: DEBUG oslo_concurrency.lockutils [req-b0a52d2e-b623-41aa-a074-6f9a1a86d913 req-d33bf36b-a966-4fd2-a41f-13766638b5c9 service nova] Releasing lock "refresh_cache-cd5a266b-617d-47c5-adb9-d161bd68f672" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.988925] env[62914]: DEBUG nova.compute.manager [req-b0a52d2e-b623-41aa-a074-6f9a1a86d913 req-d33bf36b-a966-4fd2-a41f-13766638b5c9 service nova] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Received event network-vif-deleted-21368432-56cd-4055-9388-b3ab78e2bcb3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.065651] env[62914]: DEBUG nova.network.neutron [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Successfully created port: 3cb86fc7-c6fa-4dd4-9e61-0254536a3458 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 793.139532] env[62914]: DEBUG nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 793.157799] env[62914]: INFO nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Took 0.55 seconds to detach 1 volumes for instance. [ 793.163160] env[62914]: DEBUG nova.compute.claims [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 793.163364] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.480787] env[62914]: DEBUG nova.scheduler.client.report [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.750863] env[62914]: DEBUG nova.compute.manager [req-66ee547d-9529-4eec-bb37-64cd398fbbf8 req-bdccd7a7-26ee-426f-8c04-92fa78daec7e service nova] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Received event network-changed-3cb86fc7-c6fa-4dd4-9e61-0254536a3458 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.751076] env[62914]: DEBUG nova.compute.manager [req-66ee547d-9529-4eec-bb37-64cd398fbbf8 req-bdccd7a7-26ee-426f-8c04-92fa78daec7e service nova] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Refreshing instance network info cache due to event network-changed-3cb86fc7-c6fa-4dd4-9e61-0254536a3458. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 793.751296] env[62914]: DEBUG oslo_concurrency.lockutils [req-66ee547d-9529-4eec-bb37-64cd398fbbf8 req-bdccd7a7-26ee-426f-8c04-92fa78daec7e service nova] Acquiring lock "refresh_cache-6e38d30c-2fa3-468d-8ff9-e3966333c9c8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.751439] env[62914]: DEBUG oslo_concurrency.lockutils [req-66ee547d-9529-4eec-bb37-64cd398fbbf8 req-bdccd7a7-26ee-426f-8c04-92fa78daec7e service nova] Acquired lock "refresh_cache-6e38d30c-2fa3-468d-8ff9-e3966333c9c8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.751596] env[62914]: DEBUG nova.network.neutron [req-66ee547d-9529-4eec-bb37-64cd398fbbf8 req-bdccd7a7-26ee-426f-8c04-92fa78daec7e service nova] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Refreshing network info cache for port 3cb86fc7-c6fa-4dd4-9e61-0254536a3458 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 793.985275] env[62914]: ERROR nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3cb86fc7-c6fa-4dd4-9e61-0254536a3458, please check neutron logs for more information. [ 793.985275] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 793.985275] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.985275] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 793.985275] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.985275] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 793.985275] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.985275] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 793.985275] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.985275] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 793.985275] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.985275] env[62914]: ERROR nova.compute.manager raise self.value [ 793.985275] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.985275] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 793.985275] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.985275] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 793.985762] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.985762] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 793.985762] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3cb86fc7-c6fa-4dd4-9e61-0254536a3458, please check neutron logs for more information. [ 793.985762] env[62914]: ERROR nova.compute.manager [ 793.985762] env[62914]: Traceback (most recent call last): [ 793.985762] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 793.985762] env[62914]: listener.cb(fileno) [ 793.985762] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 793.985762] env[62914]: result = function(*args, **kwargs) [ 793.985762] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 793.985762] env[62914]: return func(*args, **kwargs) [ 793.985762] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 793.985762] env[62914]: raise e [ 793.985762] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.985762] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 793.985762] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.985762] env[62914]: created_port_ids = self._update_ports_for_instance( [ 793.985762] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.985762] env[62914]: with excutils.save_and_reraise_exception(): [ 793.985762] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.985762] env[62914]: self.force_reraise() [ 793.985762] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.985762] env[62914]: raise self.value [ 793.985762] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.985762] env[62914]: updated_port = self._update_port( [ 793.985762] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.985762] env[62914]: _ensure_no_port_binding_failure(port) [ 793.985762] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.985762] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 793.986738] env[62914]: nova.exception.PortBindingFailed: Binding failed for port 3cb86fc7-c6fa-4dd4-9e61-0254536a3458, please check neutron logs for more information. [ 793.986738] env[62914]: Removing descriptor: 19 [ 793.986738] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.856s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.986835] env[62914]: ERROR nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ea8ae277-71c8-4d99-b08e-9c135e326363, please check neutron logs for more information. [ 793.986835] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Traceback (most recent call last): [ 793.986835] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 793.986835] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] self.driver.spawn(context, instance, image_meta, [ 793.986835] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 793.986835] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 793.986835] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 793.986835] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] vm_ref = self.build_virtual_machine(instance, [ 793.986835] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 793.986835] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] vif_infos = vmwarevif.get_vif_info(self._session, [ 793.986835] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 793.987171] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] for vif in network_info: [ 793.987171] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 793.987171] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] return self._sync_wrapper(fn, *args, **kwargs) [ 793.987171] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 793.987171] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] self.wait() [ 793.987171] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 793.987171] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] self[:] = self._gt.wait() [ 793.987171] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 793.987171] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] return self._exit_event.wait() [ 793.987171] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 793.987171] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] result = hub.switch() [ 793.987171] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 793.987171] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] return self.greenlet.switch() [ 793.987536] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 793.987536] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] result = function(*args, **kwargs) [ 793.987536] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 793.987536] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] return func(*args, **kwargs) [ 793.987536] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 793.987536] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] raise e [ 793.987536] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.987536] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] nwinfo = self.network_api.allocate_for_instance( [ 793.987536] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.987536] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] created_port_ids = self._update_ports_for_instance( [ 793.987536] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.987536] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] with excutils.save_and_reraise_exception(): [ 793.987536] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.987887] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] self.force_reraise() [ 793.987887] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.987887] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] raise self.value [ 793.987887] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.987887] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] updated_port = self._update_port( [ 793.987887] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.987887] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] _ensure_no_port_binding_failure(port) [ 793.987887] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.987887] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] raise exception.PortBindingFailed(port_id=port['id']) [ 793.987887] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] nova.exception.PortBindingFailed: Binding failed for port ea8ae277-71c8-4d99-b08e-9c135e326363, please check neutron logs for more information. [ 793.987887] env[62914]: ERROR nova.compute.manager [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] [ 793.988378] env[62914]: DEBUG nova.compute.utils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Binding failed for port ea8ae277-71c8-4d99-b08e-9c135e326363, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 793.989194] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.209s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.992917] env[62914]: DEBUG nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Build of instance 364c5d87-f261-439a-920e-874ee0bcf83b was re-scheduled: Binding failed for port ea8ae277-71c8-4d99-b08e-9c135e326363, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 793.993340] env[62914]: DEBUG nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 793.993561] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "refresh_cache-364c5d87-f261-439a-920e-874ee0bcf83b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.993701] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquired lock "refresh_cache-364c5d87-f261-439a-920e-874ee0bcf83b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.993852] env[62914]: DEBUG nova.network.neutron [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 794.154517] env[62914]: DEBUG nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 794.179712] env[62914]: DEBUG nova.virt.hardware [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 794.179953] env[62914]: DEBUG nova.virt.hardware [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 794.180118] env[62914]: DEBUG nova.virt.hardware [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 794.180299] env[62914]: DEBUG nova.virt.hardware [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 794.180456] env[62914]: DEBUG nova.virt.hardware [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 794.180635] env[62914]: DEBUG nova.virt.hardware [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 794.180847] env[62914]: DEBUG nova.virt.hardware [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 794.180999] env[62914]: DEBUG nova.virt.hardware [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 794.181176] env[62914]: DEBUG nova.virt.hardware [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 794.181335] env[62914]: DEBUG nova.virt.hardware [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 794.181501] env[62914]: DEBUG nova.virt.hardware [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 794.182345] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c423ad-3332-474f-a217-e95fdcf9f933 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.190098] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2830ae47-07e6-49e5-9f58-e72599acaec8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.204258] env[62914]: ERROR nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3cb86fc7-c6fa-4dd4-9e61-0254536a3458, please check neutron logs for more information. [ 794.204258] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Traceback (most recent call last): [ 794.204258] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 794.204258] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] yield resources [ 794.204258] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 794.204258] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] self.driver.spawn(context, instance, image_meta, [ 794.204258] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 794.204258] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 794.204258] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 794.204258] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] vm_ref = self.build_virtual_machine(instance, [ 794.204258] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 794.204603] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] vif_infos = vmwarevif.get_vif_info(self._session, [ 794.204603] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 794.204603] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] for vif in network_info: [ 794.204603] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 794.204603] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] return self._sync_wrapper(fn, *args, **kwargs) [ 794.204603] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 794.204603] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] self.wait() [ 794.204603] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 794.204603] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] self[:] = self._gt.wait() [ 794.204603] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 794.204603] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] return self._exit_event.wait() [ 794.204603] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 794.204603] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] current.throw(*self._exc) [ 794.205056] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 794.205056] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] result = function(*args, **kwargs) [ 794.205056] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 794.205056] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] return func(*args, **kwargs) [ 794.205056] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 794.205056] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] raise e [ 794.205056] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 794.205056] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] nwinfo = self.network_api.allocate_for_instance( [ 794.205056] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 794.205056] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] created_port_ids = self._update_ports_for_instance( [ 794.205056] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 794.205056] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] with excutils.save_and_reraise_exception(): [ 794.205056] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 794.205570] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] self.force_reraise() [ 794.205570] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 794.205570] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] raise self.value [ 794.205570] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 794.205570] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] updated_port = self._update_port( [ 794.205570] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 794.205570] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] _ensure_no_port_binding_failure(port) [ 794.205570] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 794.205570] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] raise exception.PortBindingFailed(port_id=port['id']) [ 794.205570] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] nova.exception.PortBindingFailed: Binding failed for port 3cb86fc7-c6fa-4dd4-9e61-0254536a3458, please check neutron logs for more information. [ 794.205570] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] [ 794.205570] env[62914]: INFO nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Terminating instance [ 794.206781] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "refresh_cache-6e38d30c-2fa3-468d-8ff9-e3966333c9c8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.271109] env[62914]: DEBUG nova.network.neutron [req-66ee547d-9529-4eec-bb37-64cd398fbbf8 req-bdccd7a7-26ee-426f-8c04-92fa78daec7e service nova] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.348867] env[62914]: DEBUG nova.network.neutron [req-66ee547d-9529-4eec-bb37-64cd398fbbf8 req-bdccd7a7-26ee-426f-8c04-92fa78daec7e service nova] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.512471] env[62914]: DEBUG nova.network.neutron [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.593051] env[62914]: DEBUG nova.network.neutron [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.796512] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5154fe1e-7dd4-422a-bed6-d5899043e9ba {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.804467] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296b9585-518f-4b7c-9710-2b3fff4bc9be {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.835089] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a712badf-d90a-4324-b6a7-ed6351e53a64 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.842495] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f20171c7-8699-4741-ab58-d60ab09148b0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.855878] env[62914]: DEBUG oslo_concurrency.lockutils [req-66ee547d-9529-4eec-bb37-64cd398fbbf8 req-bdccd7a7-26ee-426f-8c04-92fa78daec7e service nova] Releasing lock "refresh_cache-6e38d30c-2fa3-468d-8ff9-e3966333c9c8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.856539] env[62914]: DEBUG nova.compute.provider_tree [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.858457] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "refresh_cache-6e38d30c-2fa3-468d-8ff9-e3966333c9c8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.858637] env[62914]: DEBUG nova.network.neutron [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.098767] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Releasing lock "refresh_cache-364c5d87-f261-439a-920e-874ee0bcf83b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.098951] env[62914]: DEBUG nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 795.099148] env[62914]: DEBUG nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 795.099453] env[62914]: DEBUG nova.network.neutron [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 795.112677] env[62914]: DEBUG nova.network.neutron [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.361017] env[62914]: DEBUG nova.scheduler.client.report [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.380404] env[62914]: DEBUG nova.network.neutron [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.465094] env[62914]: DEBUG nova.network.neutron [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.615466] env[62914]: DEBUG nova.network.neutron [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.774647] env[62914]: DEBUG nova.compute.manager [req-71a49e87-02e0-44ba-9076-b6c836b1655e req-de62d480-3177-4bcd-b64a-bc6b83b70e3d service nova] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Received event network-vif-deleted-3cb86fc7-c6fa-4dd4-9e61-0254536a3458 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.867464] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.878s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.868470] env[62914]: ERROR nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6e9fb149-2aaa-4699-ac9d-371d56feba58, please check neutron logs for more information. [ 795.868470] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] Traceback (most recent call last): [ 795.868470] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 795.868470] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] self.driver.spawn(context, instance, image_meta, [ 795.868470] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 795.868470] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 795.868470] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 795.868470] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] vm_ref = self.build_virtual_machine(instance, [ 795.868470] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 795.868470] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] vif_infos = vmwarevif.get_vif_info(self._session, [ 795.868470] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 795.868841] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] for vif in network_info: [ 795.868841] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 795.868841] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] return self._sync_wrapper(fn, *args, **kwargs) [ 795.868841] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 795.868841] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] self.wait() [ 795.868841] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 795.868841] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] self[:] = self._gt.wait() [ 795.868841] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 795.868841] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] return self._exit_event.wait() [ 795.868841] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 795.868841] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] result = hub.switch() [ 795.868841] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 795.868841] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] return self.greenlet.switch() [ 795.869206] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 795.869206] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] result = function(*args, **kwargs) [ 795.869206] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 795.869206] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] return func(*args, **kwargs) [ 795.869206] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 795.869206] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] raise e [ 795.869206] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 795.869206] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] nwinfo = self.network_api.allocate_for_instance( [ 795.869206] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 795.869206] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] created_port_ids = self._update_ports_for_instance( [ 795.869206] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 795.869206] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] with excutils.save_and_reraise_exception(): [ 795.869206] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.869672] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] self.force_reraise() [ 795.869672] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.869672] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] raise self.value [ 795.869672] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 795.869672] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] updated_port = self._update_port( [ 795.869672] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 795.869672] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] _ensure_no_port_binding_failure(port) [ 795.869672] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 795.869672] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] raise exception.PortBindingFailed(port_id=port['id']) [ 795.869672] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] nova.exception.PortBindingFailed: Binding failed for port 6e9fb149-2aaa-4699-ac9d-371d56feba58, please check neutron logs for more information. [ 795.869672] env[62914]: ERROR nova.compute.manager [instance: 39489032-5430-471c-8d44-2c74682ed03b] [ 795.869973] env[62914]: DEBUG nova.compute.utils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Binding failed for port 6e9fb149-2aaa-4699-ac9d-371d56feba58, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 795.871077] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.195s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.875141] env[62914]: DEBUG nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Build of instance 39489032-5430-471c-8d44-2c74682ed03b was re-scheduled: Binding failed for port 6e9fb149-2aaa-4699-ac9d-371d56feba58, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 795.875776] env[62914]: DEBUG nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 795.876125] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "refresh_cache-39489032-5430-471c-8d44-2c74682ed03b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.876379] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquired lock "refresh_cache-39489032-5430-471c-8d44-2c74682ed03b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.876645] env[62914]: DEBUG nova.network.neutron [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.970367] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "refresh_cache-6e38d30c-2fa3-468d-8ff9-e3966333c9c8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.970783] env[62914]: DEBUG nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 795.970973] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 795.971313] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d1ac625-6969-44de-8c21-c6f44e0e9541 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.981096] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-582cd52c-7d63-4bb2-9c5c-a933cdee2ddc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.002922] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6e38d30c-2fa3-468d-8ff9-e3966333c9c8 could not be found. [ 796.002922] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 796.002922] env[62914]: INFO nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Took 0.03 seconds to destroy the instance on the hypervisor. [ 796.003187] env[62914]: DEBUG oslo.service.loopingcall [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.003451] env[62914]: DEBUG nova.compute.manager [-] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 796.003562] env[62914]: DEBUG nova.network.neutron [-] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 796.024523] env[62914]: DEBUG nova.network.neutron [-] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.118116] env[62914]: INFO nova.compute.manager [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 364c5d87-f261-439a-920e-874ee0bcf83b] Took 1.02 seconds to deallocate network for instance. [ 796.399046] env[62914]: DEBUG nova.network.neutron [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.478466] env[62914]: DEBUG nova.network.neutron [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.527079] env[62914]: DEBUG nova.network.neutron [-] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.594518] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52deb8ec-fd0d-4008-9140-20b0883ef6a9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.601444] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5689e033-af22-46dc-83cb-df9d4607df17 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.633432] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d430ab-51c3-447c-bbb6-95eeae96ede1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.640755] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94251fc4-153d-4cd2-b553-9aca915c771d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.653307] env[62914]: DEBUG nova.compute.provider_tree [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.981812] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Releasing lock "refresh_cache-39489032-5430-471c-8d44-2c74682ed03b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.981812] env[62914]: DEBUG nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 796.981812] env[62914]: DEBUG nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 796.981812] env[62914]: DEBUG nova.network.neutron [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 796.993982] env[62914]: DEBUG nova.network.neutron [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.030818] env[62914]: INFO nova.compute.manager [-] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Took 1.03 seconds to deallocate network for instance. [ 797.033098] env[62914]: DEBUG nova.compute.claims [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 797.033278] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.156041] env[62914]: INFO nova.scheduler.client.report [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Deleted allocations for instance 364c5d87-f261-439a-920e-874ee0bcf83b [ 797.162015] env[62914]: DEBUG nova.scheduler.client.report [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.497300] env[62914]: DEBUG nova.network.neutron [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.667305] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.796s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.667923] env[62914]: ERROR nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d34e9c7a-0dd0-4619-9d18-9fc611717c55, please check neutron logs for more information. [ 797.667923] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Traceback (most recent call last): [ 797.667923] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 797.667923] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] self.driver.spawn(context, instance, image_meta, [ 797.667923] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 797.667923] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 797.667923] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 797.667923] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] vm_ref = self.build_virtual_machine(instance, [ 797.667923] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 797.667923] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] vif_infos = vmwarevif.get_vif_info(self._session, [ 797.667923] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 797.668295] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] for vif in network_info: [ 797.668295] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 797.668295] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] return self._sync_wrapper(fn, *args, **kwargs) [ 797.668295] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 797.668295] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] self.wait() [ 797.668295] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 797.668295] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] self[:] = self._gt.wait() [ 797.668295] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 797.668295] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] return self._exit_event.wait() [ 797.668295] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 797.668295] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] result = hub.switch() [ 797.668295] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 797.668295] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] return self.greenlet.switch() [ 797.668613] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 797.668613] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] result = function(*args, **kwargs) [ 797.668613] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 797.668613] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] return func(*args, **kwargs) [ 797.668613] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 797.668613] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] raise e [ 797.668613] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 797.668613] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] nwinfo = self.network_api.allocate_for_instance( [ 797.668613] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 797.668613] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] created_port_ids = self._update_ports_for_instance( [ 797.668613] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 797.668613] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] with excutils.save_and_reraise_exception(): [ 797.668613] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 797.668984] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] self.force_reraise() [ 797.668984] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 797.668984] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] raise self.value [ 797.668984] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 797.668984] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] updated_port = self._update_port( [ 797.668984] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 797.668984] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] _ensure_no_port_binding_failure(port) [ 797.668984] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 797.668984] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] raise exception.PortBindingFailed(port_id=port['id']) [ 797.668984] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] nova.exception.PortBindingFailed: Binding failed for port d34e9c7a-0dd0-4619-9d18-9fc611717c55, please check neutron logs for more information. [ 797.668984] env[62914]: ERROR nova.compute.manager [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] [ 797.669244] env[62914]: DEBUG nova.compute.utils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Binding failed for port d34e9c7a-0dd0-4619-9d18-9fc611717c55, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 797.670044] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c014020-7308-4ceb-a114-50f293db4414 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "364c5d87-f261-439a-920e-874ee0bcf83b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 179.547s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.670479] env[62914]: DEBUG nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Build of instance 6201e8b7-ca3d-4eb0-8c45-67f54e66173c was re-scheduled: Binding failed for port d34e9c7a-0dd0-4619-9d18-9fc611717c55, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 797.670928] env[62914]: DEBUG nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 797.671164] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Acquiring lock "refresh_cache-6201e8b7-ca3d-4eb0-8c45-67f54e66173c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.671305] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Acquired lock "refresh_cache-6201e8b7-ca3d-4eb0-8c45-67f54e66173c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.671460] env[62914]: DEBUG nova.network.neutron [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 797.672406] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.042s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.999987] env[62914]: INFO nova.compute.manager [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: 39489032-5430-471c-8d44-2c74682ed03b] Took 1.02 seconds to deallocate network for instance. [ 798.182133] env[62914]: DEBUG nova.compute.manager [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 798.200086] env[62914]: DEBUG nova.network.neutron [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.290923] env[62914]: DEBUG nova.network.neutron [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.404091] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978c0ea6-b8ef-47e5-b18a-2614cab5518f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.411804] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b100373-c890-4561-af63-e8aab4bdaf37 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.441905] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-561e47b9-57cf-431f-a1f2-38bcc8ebd946 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.449308] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7b279a-8424-4905-a4a5-75b19cf6c3bc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.463894] env[62914]: DEBUG nova.compute.provider_tree [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.704658] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.793522] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Releasing lock "refresh_cache-6201e8b7-ca3d-4eb0-8c45-67f54e66173c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.793773] env[62914]: DEBUG nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 798.793949] env[62914]: DEBUG nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 798.794126] env[62914]: DEBUG nova.network.neutron [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 798.808775] env[62914]: DEBUG nova.network.neutron [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.966948] env[62914]: DEBUG nova.scheduler.client.report [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.036331] env[62914]: INFO nova.scheduler.client.report [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Deleted allocations for instance 39489032-5430-471c-8d44-2c74682ed03b [ 799.311640] env[62914]: DEBUG nova.network.neutron [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.472413] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.800s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.472989] env[62914]: ERROR nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7e0d56a7-513e-4ab7-b310-b8ccf72cd976, please check neutron logs for more information. [ 799.472989] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Traceback (most recent call last): [ 799.472989] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 799.472989] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] self.driver.spawn(context, instance, image_meta, [ 799.472989] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 799.472989] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] self._vmops.spawn(context, instance, image_meta, injected_files, [ 799.472989] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 799.472989] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] vm_ref = self.build_virtual_machine(instance, [ 799.472989] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 799.472989] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] vif_infos = vmwarevif.get_vif_info(self._session, [ 799.472989] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 799.473349] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] for vif in network_info: [ 799.473349] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 799.473349] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] return self._sync_wrapper(fn, *args, **kwargs) [ 799.473349] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 799.473349] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] self.wait() [ 799.473349] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 799.473349] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] self[:] = self._gt.wait() [ 799.473349] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 799.473349] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] return self._exit_event.wait() [ 799.473349] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 799.473349] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] result = hub.switch() [ 799.473349] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 799.473349] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] return self.greenlet.switch() [ 799.473640] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 799.473640] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] result = function(*args, **kwargs) [ 799.473640] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 799.473640] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] return func(*args, **kwargs) [ 799.473640] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 799.473640] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] raise e [ 799.473640] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 799.473640] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] nwinfo = self.network_api.allocate_for_instance( [ 799.473640] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 799.473640] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] created_port_ids = self._update_ports_for_instance( [ 799.473640] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 799.473640] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] with excutils.save_and_reraise_exception(): [ 799.473640] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 799.473926] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] self.force_reraise() [ 799.473926] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 799.473926] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] raise self.value [ 799.473926] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 799.473926] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] updated_port = self._update_port( [ 799.473926] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 799.473926] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] _ensure_no_port_binding_failure(port) [ 799.473926] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 799.473926] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] raise exception.PortBindingFailed(port_id=port['id']) [ 799.473926] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] nova.exception.PortBindingFailed: Binding failed for port 7e0d56a7-513e-4ab7-b310-b8ccf72cd976, please check neutron logs for more information. [ 799.473926] env[62914]: ERROR nova.compute.manager [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] [ 799.474248] env[62914]: DEBUG nova.compute.utils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Binding failed for port 7e0d56a7-513e-4ab7-b310-b8ccf72cd976, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 799.475074] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.941s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.476793] env[62914]: INFO nova.compute.claims [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.479890] env[62914]: DEBUG nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Build of instance aa1e66d1-03fb-4340-83f1-eb79c29b6934 was re-scheduled: Binding failed for port 7e0d56a7-513e-4ab7-b310-b8ccf72cd976, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 799.480372] env[62914]: DEBUG nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 799.480619] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Acquiring lock "refresh_cache-aa1e66d1-03fb-4340-83f1-eb79c29b6934" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.480775] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Acquired lock "refresh_cache-aa1e66d1-03fb-4340-83f1-eb79c29b6934" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.480942] env[62914]: DEBUG nova.network.neutron [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 799.548250] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6f21fe65-7dcd-4323-b52d-97a2a751809d tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "39489032-5430-471c-8d44-2c74682ed03b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.787s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.815429] env[62914]: INFO nova.compute.manager [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] [instance: 6201e8b7-ca3d-4eb0-8c45-67f54e66173c] Took 1.02 seconds to deallocate network for instance. [ 800.001426] env[62914]: DEBUG nova.network.neutron [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.051666] env[62914]: DEBUG nova.compute.manager [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.073243] env[62914]: DEBUG nova.network.neutron [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.311112] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.311112] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.571924] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.576718] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Releasing lock "refresh_cache-aa1e66d1-03fb-4340-83f1-eb79c29b6934" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.576718] env[62914]: DEBUG nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 800.576946] env[62914]: DEBUG nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 800.577611] env[62914]: DEBUG nova.network.neutron [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 800.602606] env[62914]: DEBUG nova.network.neutron [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.721560] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0ce063-5367-4e49-bd4c-dabcc66ba26d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.729815] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1637d53-3c66-4083-a0db-407a4ac249bc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.758928] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cea5c79-766c-4e30-b48e-7245b6dcc349 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.765895] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e8cf32-6a02-4601-947f-e2c6aff1c932 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.778893] env[62914]: DEBUG nova.compute.provider_tree [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.848271] env[62914]: INFO nova.scheduler.client.report [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Deleted allocations for instance 6201e8b7-ca3d-4eb0-8c45-67f54e66173c [ 801.105191] env[62914]: DEBUG nova.network.neutron [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.282821] env[62914]: DEBUG nova.scheduler.client.report [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 801.357037] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ce80ea29-f27b-4b59-beb1-51653690a637 tempest-ServerPasswordTestJSON-132971919 tempest-ServerPasswordTestJSON-132971919-project-member] Lock "6201e8b7-ca3d-4eb0-8c45-67f54e66173c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.580s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.607588] env[62914]: INFO nova.compute.manager [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] [instance: aa1e66d1-03fb-4340-83f1-eb79c29b6934] Took 1.03 seconds to deallocate network for instance. [ 801.790607] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.315s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.791149] env[62914]: DEBUG nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 801.793743] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.906s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.801624] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "c7cf71d7-b823-471f-949f-9ef2b539d972" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.801844] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.859607] env[62914]: DEBUG nova.compute.manager [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 802.302791] env[62914]: DEBUG nova.compute.utils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 802.308019] env[62914]: DEBUG nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 802.308019] env[62914]: DEBUG nova.network.neutron [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 802.363829] env[62914]: DEBUG nova.policy [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '004ab0e1d0484e459b97d33e736eb93b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '76aa8b10dae94f4a8ec30728965f120a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 802.381574] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.573939] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a56415-b68e-4737-a168-1d0c84b278af {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.583870] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d6519f-d158-4962-8ba2-4d2b0a8a0da5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.623256] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ebb455-772f-46e9-8cfb-f58233e0cbe9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.632751] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f97b8d4-1964-4724-a36f-c539bed07473 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.652066] env[62914]: DEBUG nova.compute.provider_tree [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.654188] env[62914]: INFO nova.scheduler.client.report [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Deleted allocations for instance aa1e66d1-03fb-4340-83f1-eb79c29b6934 [ 802.666303] env[62914]: DEBUG nova.network.neutron [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Successfully created port: f326bcf5-2bc0-42f3-ae49-60e6029eab73 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 802.812146] env[62914]: DEBUG nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 803.164045] env[62914]: DEBUG nova.scheduler.client.report [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 803.167100] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2707965f-9d78-4e2d-a418-f248a6c1421c tempest-ServerAddressesNegativeTestJSON-207972449 tempest-ServerAddressesNegativeTestJSON-207972449-project-member] Lock "aa1e66d1-03fb-4340-83f1-eb79c29b6934" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.506s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.304887] env[62914]: DEBUG nova.network.neutron [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Successfully created port: 682ccc72-1236-4a30-8250-e7853bc06945 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.669323] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.875s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.670045] env[62914]: ERROR nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9bd49c11-a193-491f-ac82-dfb5b731f817, please check neutron logs for more information. [ 803.670045] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Traceback (most recent call last): [ 803.670045] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 803.670045] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] self.driver.spawn(context, instance, image_meta, [ 803.670045] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 803.670045] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] self._vmops.spawn(context, instance, image_meta, injected_files, [ 803.670045] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 803.670045] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] vm_ref = self.build_virtual_machine(instance, [ 803.670045] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 803.670045] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] vif_infos = vmwarevif.get_vif_info(self._session, [ 803.670045] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 803.670365] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] for vif in network_info: [ 803.670365] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 803.670365] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] return self._sync_wrapper(fn, *args, **kwargs) [ 803.670365] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 803.670365] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] self.wait() [ 803.670365] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 803.670365] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] self[:] = self._gt.wait() [ 803.670365] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 803.670365] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] return self._exit_event.wait() [ 803.670365] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 803.670365] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] current.throw(*self._exc) [ 803.670365] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 803.670365] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] result = function(*args, **kwargs) [ 803.670715] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 803.670715] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] return func(*args, **kwargs) [ 803.670715] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 803.670715] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] raise e [ 803.670715] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 803.670715] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] nwinfo = self.network_api.allocate_for_instance( [ 803.670715] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 803.670715] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] created_port_ids = self._update_ports_for_instance( [ 803.670715] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 803.670715] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] with excutils.save_and_reraise_exception(): [ 803.670715] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 803.670715] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] self.force_reraise() [ 803.670715] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 803.671085] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] raise self.value [ 803.671085] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 803.671085] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] updated_port = self._update_port( [ 803.671085] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 803.671085] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] _ensure_no_port_binding_failure(port) [ 803.671085] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 803.671085] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] raise exception.PortBindingFailed(port_id=port['id']) [ 803.671085] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] nova.exception.PortBindingFailed: Binding failed for port 9bd49c11-a193-491f-ac82-dfb5b731f817, please check neutron logs for more information. [ 803.671085] env[62914]: ERROR nova.compute.manager [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] [ 803.671085] env[62914]: DEBUG nova.compute.utils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Binding failed for port 9bd49c11-a193-491f-ac82-dfb5b731f817, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 803.672598] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.454s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.681150] env[62914]: DEBUG nova.compute.manager [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 803.685019] env[62914]: DEBUG nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Build of instance 1d12cd15-22e9-4f96-8bd9-546196cb4138 was re-scheduled: Binding failed for port 9bd49c11-a193-491f-ac82-dfb5b731f817, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 803.685019] env[62914]: DEBUG nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 803.685019] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Acquiring lock "refresh_cache-1d12cd15-22e9-4f96-8bd9-546196cb4138" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.685019] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Acquired lock "refresh_cache-1d12cd15-22e9-4f96-8bd9-546196cb4138" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.685388] env[62914]: DEBUG nova.network.neutron [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 803.822092] env[62914]: DEBUG nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 803.848871] env[62914]: DEBUG nova.virt.hardware [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 803.849112] env[62914]: DEBUG nova.virt.hardware [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 803.849265] env[62914]: DEBUG nova.virt.hardware [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 803.849529] env[62914]: DEBUG nova.virt.hardware [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 803.849760] env[62914]: DEBUG nova.virt.hardware [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 803.849918] env[62914]: DEBUG nova.virt.hardware [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 803.850216] env[62914]: DEBUG nova.virt.hardware [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 803.850367] env[62914]: DEBUG nova.virt.hardware [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 803.850536] env[62914]: DEBUG nova.virt.hardware [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 803.850700] env[62914]: DEBUG nova.virt.hardware [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 803.850885] env[62914]: DEBUG nova.virt.hardware [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 803.851889] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc16e8c8-9fdd-4b5e-a5ed-3cc15c1c9ca5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.860378] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2953e16-80f5-4556-b6f0-f6ecf2b3860a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.206624] env[62914]: DEBUG nova.network.neutron [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.210210] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.270811] env[62914]: DEBUG nova.network.neutron [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.485685] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0d7634-e1ba-4e88-a3af-c3d1e81e3808 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.494563] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f0652a1-cbbb-41ad-8d17-2bd962fe3397 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.527612] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed2a2663-3117-480d-b774-cf99bab31916 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.535488] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22390765-1859-4a3a-af88-deed2bf4aab4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.549303] env[62914]: DEBUG nova.compute.provider_tree [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.712031] env[62914]: DEBUG nova.compute.manager [req-2db9f401-d64e-4b91-ba01-f8fa44ba5b36 req-8a6e810a-3760-4ea9-850b-333f1ea492dc service nova] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Received event network-changed-f326bcf5-2bc0-42f3-ae49-60e6029eab73 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.712627] env[62914]: DEBUG nova.compute.manager [req-2db9f401-d64e-4b91-ba01-f8fa44ba5b36 req-8a6e810a-3760-4ea9-850b-333f1ea492dc service nova] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Refreshing instance network info cache due to event network-changed-f326bcf5-2bc0-42f3-ae49-60e6029eab73. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 804.713018] env[62914]: DEBUG oslo_concurrency.lockutils [req-2db9f401-d64e-4b91-ba01-f8fa44ba5b36 req-8a6e810a-3760-4ea9-850b-333f1ea492dc service nova] Acquiring lock "refresh_cache-8a342650-70f0-4d5b-8086-6450dcc8d95b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.713257] env[62914]: DEBUG oslo_concurrency.lockutils [req-2db9f401-d64e-4b91-ba01-f8fa44ba5b36 req-8a6e810a-3760-4ea9-850b-333f1ea492dc service nova] Acquired lock "refresh_cache-8a342650-70f0-4d5b-8086-6450dcc8d95b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.713519] env[62914]: DEBUG nova.network.neutron [req-2db9f401-d64e-4b91-ba01-f8fa44ba5b36 req-8a6e810a-3760-4ea9-850b-333f1ea492dc service nova] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Refreshing network info cache for port f326bcf5-2bc0-42f3-ae49-60e6029eab73 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 804.776324] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Releasing lock "refresh_cache-1d12cd15-22e9-4f96-8bd9-546196cb4138" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.776324] env[62914]: DEBUG nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 804.776324] env[62914]: DEBUG nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 804.776324] env[62914]: DEBUG nova.network.neutron [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 804.803287] env[62914]: DEBUG nova.network.neutron [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.855019] env[62914]: ERROR nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f326bcf5-2bc0-42f3-ae49-60e6029eab73, please check neutron logs for more information. [ 804.855019] env[62914]: ERROR nova.compute.manager Traceback (most recent call last): [ 804.855019] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.855019] env[62914]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 804.855019] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.855019] env[62914]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 804.855019] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.855019] env[62914]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 804.855019] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.855019] env[62914]: ERROR nova.compute.manager self.force_reraise() [ 804.855019] env[62914]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.855019] env[62914]: ERROR nova.compute.manager raise self.value [ 804.855019] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.855019] env[62914]: ERROR nova.compute.manager updated_port = self._update_port( [ 804.855019] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.855019] env[62914]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 804.855533] env[62914]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.855533] env[62914]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 804.855533] env[62914]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f326bcf5-2bc0-42f3-ae49-60e6029eab73, please check neutron logs for more information. [ 804.855533] env[62914]: ERROR nova.compute.manager [ 804.855533] env[62914]: Traceback (most recent call last): [ 804.855533] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 804.855533] env[62914]: listener.cb(fileno) [ 804.855533] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 804.855533] env[62914]: result = function(*args, **kwargs) [ 804.855533] env[62914]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 804.855533] env[62914]: return func(*args, **kwargs) [ 804.855533] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 804.855533] env[62914]: raise e [ 804.855533] env[62914]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.855533] env[62914]: nwinfo = self.network_api.allocate_for_instance( [ 804.855533] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.855533] env[62914]: created_port_ids = self._update_ports_for_instance( [ 804.855533] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.855533] env[62914]: with excutils.save_and_reraise_exception(): [ 804.855533] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.855533] env[62914]: self.force_reraise() [ 804.855533] env[62914]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.855533] env[62914]: raise self.value [ 804.855533] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.855533] env[62914]: updated_port = self._update_port( [ 804.855533] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.855533] env[62914]: _ensure_no_port_binding_failure(port) [ 804.855533] env[62914]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.855533] env[62914]: raise exception.PortBindingFailed(port_id=port['id']) [ 804.856315] env[62914]: nova.exception.PortBindingFailed: Binding failed for port f326bcf5-2bc0-42f3-ae49-60e6029eab73, please check neutron logs for more information. [ 804.856315] env[62914]: Removing descriptor: 19 [ 804.856315] env[62914]: ERROR nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f326bcf5-2bc0-42f3-ae49-60e6029eab73, please check neutron logs for more information. [ 804.856315] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Traceback (most recent call last): [ 804.856315] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 804.856315] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] yield resources [ 804.856315] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 804.856315] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] self.driver.spawn(context, instance, image_meta, [ 804.856315] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 804.856315] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 804.856315] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 804.856315] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] vm_ref = self.build_virtual_machine(instance, [ 804.856646] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 804.856646] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] vif_infos = vmwarevif.get_vif_info(self._session, [ 804.856646] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 804.856646] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] for vif in network_info: [ 804.856646] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 804.856646] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] return self._sync_wrapper(fn, *args, **kwargs) [ 804.856646] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 804.856646] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] self.wait() [ 804.856646] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 804.856646] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] self[:] = self._gt.wait() [ 804.856646] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 804.856646] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] return self._exit_event.wait() [ 804.856646] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 804.856985] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] result = hub.switch() [ 804.856985] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 804.856985] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] return self.greenlet.switch() [ 804.856985] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 804.856985] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] result = function(*args, **kwargs) [ 804.856985] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 804.856985] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] return func(*args, **kwargs) [ 804.856985] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 804.856985] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] raise e [ 804.856985] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.856985] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] nwinfo = self.network_api.allocate_for_instance( [ 804.856985] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.856985] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] created_port_ids = self._update_ports_for_instance( [ 804.857342] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.857342] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] with excutils.save_and_reraise_exception(): [ 804.857342] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.857342] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] self.force_reraise() [ 804.857342] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.857342] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] raise self.value [ 804.857342] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.857342] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] updated_port = self._update_port( [ 804.857342] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.857342] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] _ensure_no_port_binding_failure(port) [ 804.857342] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.857342] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] raise exception.PortBindingFailed(port_id=port['id']) [ 804.857665] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] nova.exception.PortBindingFailed: Binding failed for port f326bcf5-2bc0-42f3-ae49-60e6029eab73, please check neutron logs for more information. [ 804.857665] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] [ 804.857665] env[62914]: INFO nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Terminating instance [ 804.864276] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Acquiring lock "refresh_cache-8a342650-70f0-4d5b-8086-6450dcc8d95b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.054037] env[62914]: DEBUG nova.scheduler.client.report [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.255925] env[62914]: DEBUG nova.network.neutron [req-2db9f401-d64e-4b91-ba01-f8fa44ba5b36 req-8a6e810a-3760-4ea9-850b-333f1ea492dc service nova] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.306483] env[62914]: DEBUG nova.network.neutron [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.435397] env[62914]: DEBUG nova.network.neutron [req-2db9f401-d64e-4b91-ba01-f8fa44ba5b36 req-8a6e810a-3760-4ea9-850b-333f1ea492dc service nova] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.558488] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.886s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.560157] env[62914]: ERROR nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 86111e45-f2b4-4aca-82fc-922d5fa3dd84, please check neutron logs for more information. [ 805.560157] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Traceback (most recent call last): [ 805.560157] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 805.560157] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] self.driver.spawn(context, instance, image_meta, [ 805.560157] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 805.560157] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] self._vmops.spawn(context, instance, image_meta, injected_files, [ 805.560157] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 805.560157] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] vm_ref = self.build_virtual_machine(instance, [ 805.560157] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 805.560157] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] vif_infos = vmwarevif.get_vif_info(self._session, [ 805.560157] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 805.560694] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] for vif in network_info: [ 805.560694] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 805.560694] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] return self._sync_wrapper(fn, *args, **kwargs) [ 805.560694] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 805.560694] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] self.wait() [ 805.560694] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 805.560694] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] self[:] = self._gt.wait() [ 805.560694] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 805.560694] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] return self._exit_event.wait() [ 805.560694] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 805.560694] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] current.throw(*self._exc) [ 805.560694] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 805.560694] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] result = function(*args, **kwargs) [ 805.561100] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 805.561100] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] return func(*args, **kwargs) [ 805.561100] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 805.561100] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] raise e [ 805.561100] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 805.561100] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] nwinfo = self.network_api.allocate_for_instance( [ 805.561100] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 805.561100] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] created_port_ids = self._update_ports_for_instance( [ 805.561100] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 805.561100] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] with excutils.save_and_reraise_exception(): [ 805.561100] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 805.561100] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] self.force_reraise() [ 805.561100] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 805.561390] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] raise self.value [ 805.561390] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 805.561390] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] updated_port = self._update_port( [ 805.561390] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 805.561390] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] _ensure_no_port_binding_failure(port) [ 805.561390] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 805.561390] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] raise exception.PortBindingFailed(port_id=port['id']) [ 805.561390] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] nova.exception.PortBindingFailed: Binding failed for port 86111e45-f2b4-4aca-82fc-922d5fa3dd84, please check neutron logs for more information. [ 805.561390] env[62914]: ERROR nova.compute.manager [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] [ 805.561390] env[62914]: DEBUG nova.compute.utils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Binding failed for port 86111e45-f2b4-4aca-82fc-922d5fa3dd84, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 805.562655] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.603s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.566614] env[62914]: DEBUG nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Build of instance dcf6b4e6-3d18-4afa-be0b-a18a58a13c08 was re-scheduled: Binding failed for port 86111e45-f2b4-4aca-82fc-922d5fa3dd84, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 805.567212] env[62914]: DEBUG nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 805.567480] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Acquiring lock "refresh_cache-dcf6b4e6-3d18-4afa-be0b-a18a58a13c08" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.567741] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Acquired lock "refresh_cache-dcf6b4e6-3d18-4afa-be0b-a18a58a13c08" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.567958] env[62914]: DEBUG nova.network.neutron [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 805.815089] env[62914]: INFO nova.compute.manager [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] [instance: 1d12cd15-22e9-4f96-8bd9-546196cb4138] Took 1.04 seconds to deallocate network for instance. [ 805.937563] env[62914]: DEBUG oslo_concurrency.lockutils [req-2db9f401-d64e-4b91-ba01-f8fa44ba5b36 req-8a6e810a-3760-4ea9-850b-333f1ea492dc service nova] Releasing lock "refresh_cache-8a342650-70f0-4d5b-8086-6450dcc8d95b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.938061] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Acquired lock "refresh_cache-8a342650-70f0-4d5b-8086-6450dcc8d95b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.938296] env[62914]: DEBUG nova.network.neutron [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 806.086307] env[62914]: DEBUG nova.network.neutron [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.171552] env[62914]: DEBUG nova.network.neutron [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.312343] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e32b4c4-acb6-410c-8001-3ac460a81b44 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.325194] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b70bf38-6ef0-4faa-a334-2d111362c64d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.355032] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e26593a-9175-4164-804f-c2e12bf86e6c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.364578] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13289f1f-0ecd-4ebb-bfc6-600ada69d3ab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.378503] env[62914]: DEBUG nova.compute.provider_tree [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.487208] env[62914]: DEBUG nova.network.neutron [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.615033] env[62914]: DEBUG nova.network.neutron [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.674009] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Releasing lock "refresh_cache-dcf6b4e6-3d18-4afa-be0b-a18a58a13c08" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.674271] env[62914]: DEBUG nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 806.674454] env[62914]: DEBUG nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.674623] env[62914]: DEBUG nova.network.neutron [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 806.698815] env[62914]: DEBUG nova.network.neutron [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.755203] env[62914]: DEBUG nova.compute.manager [req-892fd7a9-3c25-47f0-9bfd-fb3dba3431ad req-d0b5da6d-4aa1-4779-9089-f1fc8f6d3f55 service nova] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Received event network-vif-deleted-f326bcf5-2bc0-42f3-ae49-60e6029eab73 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.850966] env[62914]: INFO nova.scheduler.client.report [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Deleted allocations for instance 1d12cd15-22e9-4f96-8bd9-546196cb4138 [ 806.882024] env[62914]: DEBUG nova.scheduler.client.report [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 807.117924] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Releasing lock "refresh_cache-8a342650-70f0-4d5b-8086-6450dcc8d95b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.118070] env[62914]: DEBUG nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 807.118226] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 807.118601] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-74e61902-52a3-49d6-964a-26b5f1c4e0a4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.127747] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04c2da0b-aaaf-42fc-9a74-5a48b7dee692 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.148260] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8a342650-70f0-4d5b-8086-6450dcc8d95b could not be found. [ 807.148492] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 807.148678] env[62914]: INFO nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 807.148913] env[62914]: DEBUG oslo.service.loopingcall [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.149153] env[62914]: DEBUG nova.compute.manager [-] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 807.149245] env[62914]: DEBUG nova.network.neutron [-] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 807.187321] env[62914]: DEBUG nova.network.neutron [-] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.201432] env[62914]: DEBUG nova.network.neutron [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.359549] env[62914]: DEBUG oslo_concurrency.lockutils [None req-925fd9f5-9739-475a-9b33-ab3e0d382e8a tempest-DeleteServersAdminTestJSON-1049677667 tempest-DeleteServersAdminTestJSON-1049677667-project-member] Lock "1d12cd15-22e9-4f96-8bd9-546196cb4138" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.521s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.387049] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.825s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.387705] env[62914]: ERROR nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 21368432-56cd-4055-9388-b3ab78e2bcb3, please check neutron logs for more information. [ 807.387705] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Traceback (most recent call last): [ 807.387705] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 807.387705] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] self.driver.spawn(context, instance, image_meta, [ 807.387705] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 807.387705] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] self._vmops.spawn(context, instance, image_meta, injected_files, [ 807.387705] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 807.387705] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] vm_ref = self.build_virtual_machine(instance, [ 807.387705] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 807.387705] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] vif_infos = vmwarevif.get_vif_info(self._session, [ 807.387705] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 807.388028] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] for vif in network_info: [ 807.388028] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 807.388028] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] return self._sync_wrapper(fn, *args, **kwargs) [ 807.388028] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 807.388028] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] self.wait() [ 807.388028] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 807.388028] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] self[:] = self._gt.wait() [ 807.388028] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 807.388028] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] return self._exit_event.wait() [ 807.388028] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 807.388028] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] current.throw(*self._exc) [ 807.388028] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 807.388028] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] result = function(*args, **kwargs) [ 807.388379] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 807.388379] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] return func(*args, **kwargs) [ 807.388379] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 807.388379] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] raise e [ 807.388379] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 807.388379] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] nwinfo = self.network_api.allocate_for_instance( [ 807.388379] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 807.388379] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] created_port_ids = self._update_ports_for_instance( [ 807.388379] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 807.388379] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] with excutils.save_and_reraise_exception(): [ 807.388379] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 807.388379] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] self.force_reraise() [ 807.388379] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 807.388753] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] raise self.value [ 807.388753] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 807.388753] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] updated_port = self._update_port( [ 807.388753] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 807.388753] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] _ensure_no_port_binding_failure(port) [ 807.388753] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 807.388753] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] raise exception.PortBindingFailed(port_id=port['id']) [ 807.388753] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] nova.exception.PortBindingFailed: Binding failed for port 21368432-56cd-4055-9388-b3ab78e2bcb3, please check neutron logs for more information. [ 807.388753] env[62914]: ERROR nova.compute.manager [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] [ 807.388753] env[62914]: DEBUG nova.compute.utils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Binding failed for port 21368432-56cd-4055-9388-b3ab78e2bcb3, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 807.389647] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.226s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.392997] env[62914]: DEBUG nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Build of instance cd5a266b-617d-47c5-adb9-d161bd68f672 was re-scheduled: Binding failed for port 21368432-56cd-4055-9388-b3ab78e2bcb3, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 807.393796] env[62914]: DEBUG nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 807.394121] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Acquiring lock "refresh_cache-cd5a266b-617d-47c5-adb9-d161bd68f672" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.394316] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Acquired lock "refresh_cache-cd5a266b-617d-47c5-adb9-d161bd68f672" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.394510] env[62914]: DEBUG nova.network.neutron [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.704778] env[62914]: INFO nova.compute.manager [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] [instance: dcf6b4e6-3d18-4afa-be0b-a18a58a13c08] Took 1.03 seconds to deallocate network for instance. [ 807.861740] env[62914]: DEBUG nova.compute.manager [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 807.921130] env[62914]: DEBUG nova.network.neutron [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.009582] env[62914]: DEBUG nova.network.neutron [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.048231] env[62914]: DEBUG nova.network.neutron [-] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.093385] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ebab2e-4485-438b-8f94-fff6a5daecbf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.102277] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c69c6ef-5777-414b-9cf0-e25c4989341b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.132688] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b8f46c-324d-4bc2-9db0-24bb591201f6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.140077] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ebf491-359a-4072-9c8b-ef385772634b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.155683] env[62914]: DEBUG nova.compute.provider_tree [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.378983] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.513683] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Releasing lock "refresh_cache-cd5a266b-617d-47c5-adb9-d161bd68f672" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.513683] env[62914]: DEBUG nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 808.513683] env[62914]: DEBUG nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.513683] env[62914]: DEBUG nova.network.neutron [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 808.534939] env[62914]: DEBUG nova.network.neutron [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.553204] env[62914]: INFO nova.compute.manager [-] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Took 1.40 seconds to deallocate network for instance. [ 808.557654] env[62914]: DEBUG nova.compute.claims [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Aborting claim: {{(pid=62914) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 808.557826] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.658104] env[62914]: DEBUG nova.scheduler.client.report [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.742619] env[62914]: INFO nova.scheduler.client.report [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Deleted allocations for instance dcf6b4e6-3d18-4afa-be0b-a18a58a13c08 [ 809.037547] env[62914]: DEBUG nova.network.neutron [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.163497] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.774s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.164289] env[62914]: ERROR nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bb5cc316-1f83-4dad-8218-d3f8bdacb51f, please check neutron logs for more information. [ 809.164289] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Traceback (most recent call last): [ 809.164289] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 809.164289] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] self.driver.spawn(context, instance, image_meta, [ 809.164289] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 809.164289] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] self._vmops.spawn(context, instance, image_meta, injected_files, [ 809.164289] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 809.164289] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] vm_ref = self.build_virtual_machine(instance, [ 809.164289] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 809.164289] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] vif_infos = vmwarevif.get_vif_info(self._session, [ 809.164289] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 809.164628] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] for vif in network_info: [ 809.164628] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 809.164628] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] return self._sync_wrapper(fn, *args, **kwargs) [ 809.164628] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 809.164628] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] self.wait() [ 809.164628] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 809.164628] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] self[:] = self._gt.wait() [ 809.164628] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 809.164628] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] return self._exit_event.wait() [ 809.164628] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 809.164628] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] current.throw(*self._exc) [ 809.164628] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 809.164628] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] result = function(*args, **kwargs) [ 809.164928] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 809.164928] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] return func(*args, **kwargs) [ 809.164928] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 809.164928] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] raise e [ 809.164928] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 809.164928] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] nwinfo = self.network_api.allocate_for_instance( [ 809.164928] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 809.164928] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] created_port_ids = self._update_ports_for_instance( [ 809.164928] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 809.164928] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] with excutils.save_and_reraise_exception(): [ 809.164928] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 809.164928] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] self.force_reraise() [ 809.164928] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 809.165265] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] raise self.value [ 809.165265] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 809.165265] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] updated_port = self._update_port( [ 809.165265] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 809.165265] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] _ensure_no_port_binding_failure(port) [ 809.165265] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 809.165265] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] raise exception.PortBindingFailed(port_id=port['id']) [ 809.165265] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] nova.exception.PortBindingFailed: Binding failed for port bb5cc316-1f83-4dad-8218-d3f8bdacb51f, please check neutron logs for more information. [ 809.165265] env[62914]: ERROR nova.compute.manager [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] [ 809.165265] env[62914]: DEBUG nova.compute.utils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Binding failed for port bb5cc316-1f83-4dad-8218-d3f8bdacb51f, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 809.166382] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.133s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.169398] env[62914]: DEBUG nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Build of instance cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206 was re-scheduled: Binding failed for port bb5cc316-1f83-4dad-8218-d3f8bdacb51f, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 809.169838] env[62914]: DEBUG nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 809.170153] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Acquiring lock "refresh_cache-cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.170327] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Acquired lock "refresh_cache-cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.170490] env[62914]: DEBUG nova.network.neutron [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 809.255473] env[62914]: DEBUG oslo_concurrency.lockutils [None req-acd39e59-6b21-4273-b942-6a0703a8cc98 tempest-ServerActionsTestOtherB-56943281 tempest-ServerActionsTestOtherB-56943281-project-member] Lock "dcf6b4e6-3d18-4afa-be0b-a18a58a13c08" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.891s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.541637] env[62914]: INFO nova.compute.manager [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] [instance: cd5a266b-617d-47c5-adb9-d161bd68f672] Took 1.03 seconds to deallocate network for instance. [ 809.695915] env[62914]: DEBUG nova.network.neutron [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.761604] env[62914]: DEBUG nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.807118] env[62914]: DEBUG nova.network.neutron [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.946948] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5735e206-3143-4171-a248-3d755dcb6895 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.955241] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c60d68eb-86d8-4311-bb79-015a2ab3e338 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.992574] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e35c7d2-285a-48b5-92e7-c10f41d22041 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.000765] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf9f319e-e2a0-4c36-91e1-442daf3c6550 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.014934] env[62914]: DEBUG nova.compute.provider_tree [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.289971] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.309807] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Releasing lock "refresh_cache-cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.310061] env[62914]: DEBUG nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 810.310255] env[62914]: DEBUG nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 810.310422] env[62914]: DEBUG nova.network.neutron [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 810.332149] env[62914]: DEBUG nova.network.neutron [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 810.522187] env[62914]: DEBUG nova.scheduler.client.report [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 810.596780] env[62914]: INFO nova.scheduler.client.report [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Deleted allocations for instance cd5a266b-617d-47c5-adb9-d161bd68f672 [ 810.836192] env[62914]: DEBUG nova.network.neutron [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.031196] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.865s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.031816] env[62914]: ERROR nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3cb86fc7-c6fa-4dd4-9e61-0254536a3458, please check neutron logs for more information. [ 811.031816] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Traceback (most recent call last): [ 811.031816] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 811.031816] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] self.driver.spawn(context, instance, image_meta, [ 811.031816] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 811.031816] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 811.031816] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 811.031816] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] vm_ref = self.build_virtual_machine(instance, [ 811.031816] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 811.031816] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] vif_infos = vmwarevif.get_vif_info(self._session, [ 811.031816] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 811.032152] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] for vif in network_info: [ 811.032152] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 811.032152] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] return self._sync_wrapper(fn, *args, **kwargs) [ 811.032152] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 811.032152] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] self.wait() [ 811.032152] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 811.032152] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] self[:] = self._gt.wait() [ 811.032152] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 811.032152] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] return self._exit_event.wait() [ 811.032152] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 811.032152] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] current.throw(*self._exc) [ 811.032152] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 811.032152] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] result = function(*args, **kwargs) [ 811.032492] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 811.032492] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] return func(*args, **kwargs) [ 811.032492] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 811.032492] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] raise e [ 811.032492] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 811.032492] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] nwinfo = self.network_api.allocate_for_instance( [ 811.032492] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 811.032492] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] created_port_ids = self._update_ports_for_instance( [ 811.032492] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 811.032492] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] with excutils.save_and_reraise_exception(): [ 811.032492] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 811.032492] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] self.force_reraise() [ 811.032492] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 811.032770] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] raise self.value [ 811.032770] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 811.032770] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] updated_port = self._update_port( [ 811.032770] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 811.032770] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] _ensure_no_port_binding_failure(port) [ 811.032770] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 811.032770] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] raise exception.PortBindingFailed(port_id=port['id']) [ 811.032770] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] nova.exception.PortBindingFailed: Binding failed for port 3cb86fc7-c6fa-4dd4-9e61-0254536a3458, please check neutron logs for more information. [ 811.032770] env[62914]: ERROR nova.compute.manager [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] [ 811.032770] env[62914]: DEBUG nova.compute.utils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Binding failed for port 3cb86fc7-c6fa-4dd4-9e61-0254536a3458, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 811.033774] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.329s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.035680] env[62914]: INFO nova.compute.claims [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.038790] env[62914]: DEBUG nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Build of instance 6e38d30c-2fa3-468d-8ff9-e3966333c9c8 was re-scheduled: Binding failed for port 3cb86fc7-c6fa-4dd4-9e61-0254536a3458, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 811.039226] env[62914]: DEBUG nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 811.039448] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "refresh_cache-6e38d30c-2fa3-468d-8ff9-e3966333c9c8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.039619] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "refresh_cache-6e38d30c-2fa3-468d-8ff9-e3966333c9c8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.039778] env[62914]: DEBUG nova.network.neutron [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 811.108889] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b37b8d7-0ba6-4e5f-bf9a-8b60fcc3407d tempest-ServerRescueTestJSON-634319772 tempest-ServerRescueTestJSON-634319772-project-member] Lock "cd5a266b-617d-47c5-adb9-d161bd68f672" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.537s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.339850] env[62914]: INFO nova.compute.manager [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] [instance: cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206] Took 1.03 seconds to deallocate network for instance. [ 811.570291] env[62914]: DEBUG nova.network.neutron [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 811.611933] env[62914]: DEBUG nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 811.727356] env[62914]: DEBUG nova.network.neutron [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.147991] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.229900] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "refresh_cache-6e38d30c-2fa3-468d-8ff9-e3966333c9c8" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.230150] env[62914]: DEBUG nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 812.230328] env[62914]: DEBUG nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 812.230504] env[62914]: DEBUG nova.network.neutron [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 812.251887] env[62914]: DEBUG nova.network.neutron [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 812.322013] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc8d0a3-7158-4bc8-8362-1e4413f129b7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.330103] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a116be49-0418-453f-b637-2c7d0610f5d8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.365462] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec74bb3-086e-48ad-8c7e-ebc0b6264a5e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.373633] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-035282c2-2a82-4662-aec5-3b6d18da0407 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.389472] env[62914]: DEBUG nova.compute.provider_tree [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.391773] env[62914]: INFO nova.scheduler.client.report [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Deleted allocations for instance cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206 [ 812.757473] env[62914]: DEBUG nova.network.neutron [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.900966] env[62914]: DEBUG nova.scheduler.client.report [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.908100] env[62914]: DEBUG oslo_concurrency.lockutils [None req-29a0efc3-ad0f-41ba-8687-d9a845bc6275 tempest-ServersTestBootFromVolume-1913954265 tempest-ServersTestBootFromVolume-1913954265-project-member] Lock "cdd7a72a-a310-46e4-8fb5-3d2e8f0c8206" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.962s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.260738] env[62914]: INFO nova.compute.manager [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 6e38d30c-2fa3-468d-8ff9-e3966333c9c8] Took 1.03 seconds to deallocate network for instance. [ 813.407105] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.407934] env[62914]: DEBUG nova.compute.manager [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 813.410452] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.839s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.413023] env[62914]: INFO nova.compute.claims [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.415964] env[62914]: DEBUG nova.compute.manager [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 813.921453] env[62914]: DEBUG nova.compute.utils [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 813.928279] env[62914]: DEBUG nova.compute.manager [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Not allocating networking since 'none' was specified. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 813.955564] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.297739] env[62914]: INFO nova.scheduler.client.report [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Deleted allocations for instance 6e38d30c-2fa3-468d-8ff9-e3966333c9c8 [ 814.429388] env[62914]: DEBUG nova.compute.manager [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 814.673481] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86cca71f-fe89-44bc-84b9-f54523b38955 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.681254] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18794e97-e620-4625-9581-ca6c8759461a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.711358] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fdccb5-3e4e-42c4-b8ef-a15effdf67cd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.719022] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e37296-7eb8-4e96-8464-41f6da5f0a53 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.733286] env[62914]: DEBUG nova.compute.provider_tree [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.808321] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e812ead8-ec77-4db8-b270-d63f5582652e tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "6e38d30c-2fa3-468d-8ff9-e3966333c9c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.232s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.237970] env[62914]: DEBUG nova.scheduler.client.report [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.311602] env[62914]: DEBUG nova.compute.manager [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 815.441867] env[62914]: DEBUG nova.compute.manager [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 815.479449] env[62914]: DEBUG nova.virt.hardware [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 815.479740] env[62914]: DEBUG nova.virt.hardware [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 815.479940] env[62914]: DEBUG nova.virt.hardware [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 815.480151] env[62914]: DEBUG nova.virt.hardware [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 815.480298] env[62914]: DEBUG nova.virt.hardware [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 815.480443] env[62914]: DEBUG nova.virt.hardware [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 815.480646] env[62914]: DEBUG nova.virt.hardware [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 815.480800] env[62914]: DEBUG nova.virt.hardware [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 815.481148] env[62914]: DEBUG nova.virt.hardware [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 815.481148] env[62914]: DEBUG nova.virt.hardware [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 815.481284] env[62914]: DEBUG nova.virt.hardware [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 815.482366] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a2693da-dade-4224-8cb4-c9d0b82f7c9e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.491286] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b49ab3-3372-49c6-84e0-536a37bd890e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.508725] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Instance VIF info [] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 815.515513] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Creating folder: Project (e160353c3cdd4e18ac115a18f91d6727). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.515958] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c67ad9f1-7bcc-40cf-9917-8e8d3ff9581a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.531120] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Created folder: Project (e160353c3cdd4e18ac115a18f91d6727) in parent group-v288131. [ 815.531120] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Creating folder: Instances. Parent ref: group-v288160. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.531120] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-de8de6cd-1849-403c-9991-eaea71784e56 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.540105] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Created folder: Instances in parent group-v288160. [ 815.540449] env[62914]: DEBUG oslo.service.loopingcall [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.540790] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 815.541211] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-87959002-9a93-4aea-a96a-cc3a0027bee1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.563925] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 815.563925] env[62914]: value = "task-1352435" [ 815.563925] env[62914]: _type = "Task" [ 815.563925] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.574185] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352435, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.745009] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.334s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.745575] env[62914]: DEBUG nova.compute.manager [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 815.748935] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.367s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.750479] env[62914]: INFO nova.compute.claims [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 815.843298] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.079184] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352435, 'name': CreateVM_Task, 'duration_secs': 0.289477} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.079184] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 816.079184] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.079184] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.079184] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 816.079184] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff790392-8f38-439c-8e63-a82f6c480740 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.082192] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 816.082192] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]528047ba-4e00-11bf-63b4-5063d1bcfa2d" [ 816.082192] env[62914]: _type = "Task" [ 816.082192] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.091655] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]528047ba-4e00-11bf-63b4-5063d1bcfa2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.256375] env[62914]: DEBUG nova.compute.utils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 816.261462] env[62914]: DEBUG nova.compute.manager [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 816.261462] env[62914]: DEBUG nova.network.neutron [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 816.360479] env[62914]: DEBUG nova.policy [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '46e0df5386e746a7910ba8b47fbf0005', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9cb17042d6ab4ab4824e39bf809960da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 816.600020] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]528047ba-4e00-11bf-63b4-5063d1bcfa2d, 'name': SearchDatastore_Task, 'duration_secs': 0.012223} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.600357] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.600584] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 816.600813] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.600957] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.601171] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 816.601446] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a3a9e6df-01dd-4671-a664-9bca885b2e18 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.614468] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 816.614468] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 816.615616] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd702490-2bec-4cb3-9e67-ec2c92d9be4b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.621823] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 816.621823] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]524d9835-5b63-87b3-b98d-96f6541100f1" [ 816.621823] env[62914]: _type = "Task" [ 816.621823] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.634917] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]524d9835-5b63-87b3-b98d-96f6541100f1, 'name': SearchDatastore_Task, 'duration_secs': 0.008931} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.635941] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3953a19d-b6fd-4a4d-a271-5e6e06644de5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.642194] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 816.642194] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]521a27bc-4472-1bed-5e47-c4c058f5f2da" [ 816.642194] env[62914]: _type = "Task" [ 816.642194] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.657076] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]521a27bc-4472-1bed-5e47-c4c058f5f2da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.762122] env[62914]: DEBUG nova.compute.manager [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 816.964745] env[62914]: DEBUG nova.network.neutron [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Successfully created port: 299c1132-e4a5-4803-8e8d-1478b5928463 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.034624] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d544730-e7a7-4b8c-b5fe-1944907c332f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.044097] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6141db8-5361-47c1-ad62-7c748a5a4a26 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.083262] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da26042b-e24a-44a1-b5ab-d1f2084317f9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.087438] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "4bf76ce0-b5d2-4184-888c-d0ef39878356" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.087654] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "4bf76ce0-b5d2-4184-888c-d0ef39878356" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.092670] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc48d3ee-4211-4f27-8b9f-61ea4fcb41ee {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.109143] env[62914]: DEBUG nova.compute.provider_tree [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.151832] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]521a27bc-4472-1bed-5e47-c4c058f5f2da, 'name': SearchDatastore_Task, 'duration_secs': 0.010799} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.152358] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.152425] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 87ef6b06-f699-4de3-8b89-854717074406/87ef6b06-f699-4de3-8b89-854717074406.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 817.154192] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a590fbc9-ad15-451b-af31-02919b1d8e0b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.159420] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 817.159420] env[62914]: value = "task-1352436" [ 817.159420] env[62914]: _type = "Task" [ 817.159420] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.170704] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352436, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.193527] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.193642] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.614024] env[62914]: DEBUG nova.scheduler.client.report [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 817.669756] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352436, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491162} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.670053] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 87ef6b06-f699-4de3-8b89-854717074406/87ef6b06-f699-4de3-8b89-854717074406.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 817.670293] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 817.670542] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09f74029-a1a9-4233-8faa-abaecb9b0be6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.677791] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 817.677791] env[62914]: value = "task-1352437" [ 817.677791] env[62914]: _type = "Task" [ 817.677791] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.689871] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352437, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.779460] env[62914]: DEBUG nova.compute.manager [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 817.821081] env[62914]: DEBUG nova.virt.hardware [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.821290] env[62914]: DEBUG nova.virt.hardware [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.821454] env[62914]: DEBUG nova.virt.hardware [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.821632] env[62914]: DEBUG nova.virt.hardware [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.821765] env[62914]: DEBUG nova.virt.hardware [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.821938] env[62914]: DEBUG nova.virt.hardware [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.822118] env[62914]: DEBUG nova.virt.hardware [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.822278] env[62914]: DEBUG nova.virt.hardware [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.822439] env[62914]: DEBUG nova.virt.hardware [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.822592] env[62914]: DEBUG nova.virt.hardware [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.822761] env[62914]: DEBUG nova.virt.hardware [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.823768] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b01512-9d0f-4e16-8bca-6671a6d373b0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.833538] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c47e26-18a1-4754-9153-b6e51c28c1e3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.123328] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.123328] env[62914]: DEBUG nova.compute.manager [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 818.124666] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.915s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.127301] env[62914]: INFO nova.compute.claims [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 818.192017] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352437, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071908} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.192619] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 818.193453] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13544c50-86bb-4b9d-80c9-aa1f19ded467 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.214828] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 87ef6b06-f699-4de3-8b89-854717074406/87ef6b06-f699-4de3-8b89-854717074406.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 818.215378] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb577ac7-3fb5-498e-8af1-f6aa8cabe60f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.235913] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 818.235913] env[62914]: value = "task-1352438" [ 818.235913] env[62914]: _type = "Task" [ 818.235913] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.244048] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352438, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.369851] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquiring lock "29680cfb-8df3-4ff7-b6ee-0982577cc9cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.369851] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Lock "29680cfb-8df3-4ff7-b6ee-0982577cc9cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.638064] env[62914]: DEBUG nova.compute.utils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 818.639246] env[62914]: DEBUG nova.compute.manager [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 818.639404] env[62914]: DEBUG nova.network.neutron [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 818.698602] env[62914]: DEBUG nova.policy [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80089586cfda4cd2bdfc24d9f556239c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09d51090ac5d41fc966c7e132f288824', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 818.754545] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352438, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.772092] env[62914]: DEBUG nova.compute.manager [req-cfaeefc9-d318-4069-aa2a-1d2e34c4a5ce req-a5e44e80-2702-4d4e-bfc8-a293882b7a18 service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Received event network-vif-plugged-299c1132-e4a5-4803-8e8d-1478b5928463 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.772211] env[62914]: DEBUG oslo_concurrency.lockutils [req-cfaeefc9-d318-4069-aa2a-1d2e34c4a5ce req-a5e44e80-2702-4d4e-bfc8-a293882b7a18 service nova] Acquiring lock "9fab500c-344c-46cd-b060-c8f1aa0f1cba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.772416] env[62914]: DEBUG oslo_concurrency.lockutils [req-cfaeefc9-d318-4069-aa2a-1d2e34c4a5ce req-a5e44e80-2702-4d4e-bfc8-a293882b7a18 service nova] Lock "9fab500c-344c-46cd-b060-c8f1aa0f1cba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.772582] env[62914]: DEBUG oslo_concurrency.lockutils [req-cfaeefc9-d318-4069-aa2a-1d2e34c4a5ce req-a5e44e80-2702-4d4e-bfc8-a293882b7a18 service nova] Lock "9fab500c-344c-46cd-b060-c8f1aa0f1cba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.772739] env[62914]: DEBUG nova.compute.manager [req-cfaeefc9-d318-4069-aa2a-1d2e34c4a5ce req-a5e44e80-2702-4d4e-bfc8-a293882b7a18 service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] No waiting events found dispatching network-vif-plugged-299c1132-e4a5-4803-8e8d-1478b5928463 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 818.772897] env[62914]: WARNING nova.compute.manager [req-cfaeefc9-d318-4069-aa2a-1d2e34c4a5ce req-a5e44e80-2702-4d4e-bfc8-a293882b7a18 service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Received unexpected event network-vif-plugged-299c1132-e4a5-4803-8e8d-1478b5928463 for instance with vm_state building and task_state spawning. [ 818.888144] env[62914]: DEBUG nova.network.neutron [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Successfully updated port: 299c1132-e4a5-4803-8e8d-1478b5928463 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 819.137116] env[62914]: DEBUG nova.network.neutron [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Successfully created port: b12a55b6-c396-44ab-a9ef-b183e97189e3 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 819.144746] env[62914]: DEBUG nova.compute.manager [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 819.247570] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352438, 'name': ReconfigVM_Task, 'duration_secs': 0.537545} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.248027] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 87ef6b06-f699-4de3-8b89-854717074406/87ef6b06-f699-4de3-8b89-854717074406.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.249394] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee6dcc0c-02cc-4a34-a28a-32884d00506a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.259860] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 819.259860] env[62914]: value = "task-1352439" [ 819.259860] env[62914]: _type = "Task" [ 819.259860] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.268656] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352439, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.391990] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Acquiring lock "refresh_cache-9fab500c-344c-46cd-b060-c8f1aa0f1cba" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.392424] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Acquired lock "refresh_cache-9fab500c-344c-46cd-b060-c8f1aa0f1cba" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.392578] env[62914]: DEBUG nova.network.neutron [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 819.431292] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3affd1c2-8385-40a0-bd79-dacb7b96741d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.439420] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d833b8e0-abd4-433b-b059-f17eac12a44a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.470330] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7b98d3-61db-4f63-9f75-1d12cae07910 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.477801] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d03245-2fbc-4376-bf18-09a5e52ada8d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.491226] env[62914]: DEBUG nova.compute.provider_tree [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.771451] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352439, 'name': Rename_Task, 'duration_secs': 0.159876} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.771451] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 819.771451] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3eafeeac-3f8e-476b-bcc5-dc6ab783657d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.778020] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 819.778020] env[62914]: value = "task-1352440" [ 819.778020] env[62914]: _type = "Task" [ 819.778020] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.788267] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352440, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.924014] env[62914]: DEBUG nova.network.neutron [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.995622] env[62914]: DEBUG nova.scheduler.client.report [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.154406] env[62914]: DEBUG nova.compute.manager [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 820.157471] env[62914]: DEBUG nova.network.neutron [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Updating instance_info_cache with network_info: [{"id": "299c1132-e4a5-4803-8e8d-1478b5928463", "address": "fa:16:3e:d2:24:e5", "network": {"id": "3f1a31ed-b39a-473d-a7f4-979cde7098b0", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-677647149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb17042d6ab4ab4824e39bf809960da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0dd3c126-9d86-4f9a-b81c-e9627c7a5401", "external-id": "nsx-vlan-transportzone-24", "segmentation_id": 24, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap299c1132-e4", "ovs_interfaceid": "299c1132-e4a5-4803-8e8d-1478b5928463", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.179092] env[62914]: DEBUG nova.virt.hardware [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 820.179447] env[62914]: DEBUG nova.virt.hardware [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 820.179670] env[62914]: DEBUG nova.virt.hardware [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.179946] env[62914]: DEBUG nova.virt.hardware [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 820.180167] env[62914]: DEBUG nova.virt.hardware [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.180408] env[62914]: DEBUG nova.virt.hardware [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 820.180626] env[62914]: DEBUG nova.virt.hardware [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 820.180824] env[62914]: DEBUG nova.virt.hardware [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 820.181049] env[62914]: DEBUG nova.virt.hardware [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 820.181262] env[62914]: DEBUG nova.virt.hardware [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 820.181530] env[62914]: DEBUG nova.virt.hardware [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.183362] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b2271e-7789-49c2-a8f0-11be8b0e1fa7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.193784] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a22856-0e15-4c6e-88c9-68756ccb3f3c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.286356] env[62914]: DEBUG oslo_vmware.api [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352440, 'name': PowerOnVM_Task, 'duration_secs': 0.420513} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.286599] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 820.286792] env[62914]: INFO nova.compute.manager [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Took 4.85 seconds to spawn the instance on the hypervisor. [ 820.287077] env[62914]: DEBUG nova.compute.manager [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.287744] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee98f3e-f442-480a-a046-039216e51f89 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.501800] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.377s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.502346] env[62914]: DEBUG nova.compute.manager [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 820.504949] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.126s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.506359] env[62914]: INFO nova.compute.claims [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 820.639850] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Acquiring lock "f80ee46d-9209-4b60-9e5a-f7c8764f4ccb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.640117] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Lock "f80ee46d-9209-4b60-9e5a-f7c8764f4ccb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.664159] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Releasing lock "refresh_cache-9fab500c-344c-46cd-b060-c8f1aa0f1cba" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.664159] env[62914]: DEBUG nova.compute.manager [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Instance network_info: |[{"id": "299c1132-e4a5-4803-8e8d-1478b5928463", "address": "fa:16:3e:d2:24:e5", "network": {"id": "3f1a31ed-b39a-473d-a7f4-979cde7098b0", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-677647149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb17042d6ab4ab4824e39bf809960da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0dd3c126-9d86-4f9a-b81c-e9627c7a5401", "external-id": "nsx-vlan-transportzone-24", "segmentation_id": 24, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap299c1132-e4", "ovs_interfaceid": "299c1132-e4a5-4803-8e8d-1478b5928463", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 820.664290] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:24:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0dd3c126-9d86-4f9a-b81c-e9627c7a5401', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '299c1132-e4a5-4803-8e8d-1478b5928463', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 820.670795] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Creating folder: Project (9cb17042d6ab4ab4824e39bf809960da). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.671622] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ffd7247c-8be5-414a-8235-107baa93a31d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.685316] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Created folder: Project (9cb17042d6ab4ab4824e39bf809960da) in parent group-v288131. [ 820.685316] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Creating folder: Instances. Parent ref: group-v288163. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.685316] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b650259c-a4b1-48ec-9122-6a4255930cfc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.693850] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Created folder: Instances in parent group-v288163. [ 820.694092] env[62914]: DEBUG oslo.service.loopingcall [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.694437] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 820.694692] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-859e5d48-7b87-4c9a-bbdd-fbd44089b149 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.718685] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 820.718685] env[62914]: value = "task-1352443" [ 820.718685] env[62914]: _type = "Task" [ 820.718685] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.724220] env[62914]: DEBUG nova.compute.manager [req-faa2bfd4-a013-40f0-b702-4bee144980f1 req-3f1bebfb-6e95-44d8-b698-43b605b79c1e service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Received event network-vif-plugged-b12a55b6-c396-44ab-a9ef-b183e97189e3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.724333] env[62914]: DEBUG oslo_concurrency.lockutils [req-faa2bfd4-a013-40f0-b702-4bee144980f1 req-3f1bebfb-6e95-44d8-b698-43b605b79c1e service nova] Acquiring lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.725220] env[62914]: DEBUG oslo_concurrency.lockutils [req-faa2bfd4-a013-40f0-b702-4bee144980f1 req-3f1bebfb-6e95-44d8-b698-43b605b79c1e service nova] Lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.725220] env[62914]: DEBUG oslo_concurrency.lockutils [req-faa2bfd4-a013-40f0-b702-4bee144980f1 req-3f1bebfb-6e95-44d8-b698-43b605b79c1e service nova] Lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.725220] env[62914]: DEBUG nova.compute.manager [req-faa2bfd4-a013-40f0-b702-4bee144980f1 req-3f1bebfb-6e95-44d8-b698-43b605b79c1e service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] No waiting events found dispatching network-vif-plugged-b12a55b6-c396-44ab-a9ef-b183e97189e3 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.725220] env[62914]: WARNING nova.compute.manager [req-faa2bfd4-a013-40f0-b702-4bee144980f1 req-3f1bebfb-6e95-44d8-b698-43b605b79c1e service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Received unexpected event network-vif-plugged-b12a55b6-c396-44ab-a9ef-b183e97189e3 for instance with vm_state building and task_state spawning. [ 820.730434] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352443, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.805118] env[62914]: DEBUG nova.compute.manager [req-9c2d5ade-9da1-4aa7-892f-5449b1a076ff req-66b9d476-2c32-469b-83c5-280265435247 service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Received event network-changed-299c1132-e4a5-4803-8e8d-1478b5928463 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.805744] env[62914]: DEBUG nova.compute.manager [req-9c2d5ade-9da1-4aa7-892f-5449b1a076ff req-66b9d476-2c32-469b-83c5-280265435247 service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Refreshing instance network info cache due to event network-changed-299c1132-e4a5-4803-8e8d-1478b5928463. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 820.806384] env[62914]: DEBUG oslo_concurrency.lockutils [req-9c2d5ade-9da1-4aa7-892f-5449b1a076ff req-66b9d476-2c32-469b-83c5-280265435247 service nova] Acquiring lock "refresh_cache-9fab500c-344c-46cd-b060-c8f1aa0f1cba" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.806694] env[62914]: DEBUG oslo_concurrency.lockutils [req-9c2d5ade-9da1-4aa7-892f-5449b1a076ff req-66b9d476-2c32-469b-83c5-280265435247 service nova] Acquired lock "refresh_cache-9fab500c-344c-46cd-b060-c8f1aa0f1cba" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.807053] env[62914]: DEBUG nova.network.neutron [req-9c2d5ade-9da1-4aa7-892f-5449b1a076ff req-66b9d476-2c32-469b-83c5-280265435247 service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Refreshing network info cache for port 299c1132-e4a5-4803-8e8d-1478b5928463 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 820.811141] env[62914]: INFO nova.compute.manager [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Took 22.12 seconds to build instance. [ 820.834553] env[62914]: DEBUG nova.network.neutron [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Successfully updated port: b12a55b6-c396-44ab-a9ef-b183e97189e3 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 821.010355] env[62914]: DEBUG nova.compute.utils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 821.013919] env[62914]: DEBUG nova.compute.manager [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 821.014185] env[62914]: DEBUG nova.network.neutron [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 821.069904] env[62914]: DEBUG nova.policy [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2c21c59e39b94deea6c3701fe971198a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1dcda44f0d6a455bb17966e7c5061190', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 821.229608] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352443, 'name': CreateVM_Task, 'duration_secs': 0.317263} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.229608] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 821.240462] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.240558] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.240886] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 821.241149] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fbac70e-aa83-4cb0-9fae-e6efbc992f2b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.246173] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Waiting for the task: (returnval){ [ 821.246173] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ef30d3-a3a5-e241-fdea-97dac12d944c" [ 821.246173] env[62914]: _type = "Task" [ 821.246173] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.254254] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ef30d3-a3a5-e241-fdea-97dac12d944c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.315435] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82a7fe9d-0be6-4c47-aea0-205599b00501 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Lock "87ef6b06-f699-4de3-8b89-854717074406" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.785s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.339622] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.339754] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.339996] env[62914]: DEBUG nova.network.neutron [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 821.384563] env[62914]: DEBUG nova.network.neutron [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Successfully created port: b82c2ddd-9dd3-4bd8-ad01-68787d570158 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 821.514605] env[62914]: DEBUG nova.compute.manager [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 821.535828] env[62914]: DEBUG nova.network.neutron [req-9c2d5ade-9da1-4aa7-892f-5449b1a076ff req-66b9d476-2c32-469b-83c5-280265435247 service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Updated VIF entry in instance network info cache for port 299c1132-e4a5-4803-8e8d-1478b5928463. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 821.536086] env[62914]: DEBUG nova.network.neutron [req-9c2d5ade-9da1-4aa7-892f-5449b1a076ff req-66b9d476-2c32-469b-83c5-280265435247 service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Updating instance_info_cache with network_info: [{"id": "299c1132-e4a5-4803-8e8d-1478b5928463", "address": "fa:16:3e:d2:24:e5", "network": {"id": "3f1a31ed-b39a-473d-a7f4-979cde7098b0", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-677647149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb17042d6ab4ab4824e39bf809960da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0dd3c126-9d86-4f9a-b81c-e9627c7a5401", "external-id": "nsx-vlan-transportzone-24", "segmentation_id": 24, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap299c1132-e4", "ovs_interfaceid": "299c1132-e4a5-4803-8e8d-1478b5928463", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.652658] env[62914]: INFO nova.compute.manager [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Rebuilding instance [ 821.694203] env[62914]: DEBUG nova.compute.manager [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.695108] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a327a46b-fb1a-43e3-a528-8c23dfbfa09f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.758971] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ef30d3-a3a5-e241-fdea-97dac12d944c, 'name': SearchDatastore_Task, 'duration_secs': 0.008891} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.759784] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.759784] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 821.759784] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.760093] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.760093] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 821.760486] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70e0f6d5-99ca-4559-94e0-09d7055b0964 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.771680] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 821.771866] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 821.772587] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e29ad65a-d910-4eeb-8940-7b22bd29a09f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.776286] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe61fe7-adfe-4718-9c56-b60194049125 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.782344] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Waiting for the task: (returnval){ [ 821.782344] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a04c91-9641-f9cb-de2a-185b070c7c08" [ 821.782344] env[62914]: _type = "Task" [ 821.782344] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.787644] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07c9ff3-06ab-4137-96b4-3b98c68c1a91 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.795948] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a04c91-9641-f9cb-de2a-185b070c7c08, 'name': SearchDatastore_Task, 'duration_secs': 0.008445} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.821431] env[62914]: DEBUG nova.compute.manager [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 821.824245] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9da0b9ae-8b13-4106-9c79-91cdf5fdcdf6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.826989] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-537f4591-9bf0-418a-a6e9-0275b7d4b31b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.836452] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Waiting for the task: (returnval){ [ 821.836452] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]529a27c2-a3d0-4ff8-13e9-c91353162924" [ 821.836452] env[62914]: _type = "Task" [ 821.836452] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.837106] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cfe2dc2-6389-405d-8eb1-b1b4b3043b33 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.855080] env[62914]: DEBUG nova.compute.provider_tree [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.859451] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]529a27c2-a3d0-4ff8-13e9-c91353162924, 'name': SearchDatastore_Task, 'duration_secs': 0.00861} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.860035] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.860142] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 9fab500c-344c-46cd-b060-c8f1aa0f1cba/9fab500c-344c-46cd-b060-c8f1aa0f1cba.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 821.860592] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5841314f-a064-4e5b-acda-54fefc624666 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.867145] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Waiting for the task: (returnval){ [ 821.867145] env[62914]: value = "task-1352444" [ 821.867145] env[62914]: _type = "Task" [ 821.867145] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.875038] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352444, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.888197] env[62914]: DEBUG nova.network.neutron [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.022357] env[62914]: DEBUG nova.network.neutron [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Updating instance_info_cache with network_info: [{"id": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "address": "fa:16:3e:7a:f0:fa", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb12a55b6-c3", "ovs_interfaceid": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.041716] env[62914]: DEBUG oslo_concurrency.lockutils [req-9c2d5ade-9da1-4aa7-892f-5449b1a076ff req-66b9d476-2c32-469b-83c5-280265435247 service nova] Releasing lock "refresh_cache-9fab500c-344c-46cd-b060-c8f1aa0f1cba" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.208490] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 822.208871] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d07e49ba-e0e0-4f71-8923-0be68db7fc46 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.217077] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 822.217077] env[62914]: value = "task-1352445" [ 822.217077] env[62914]: _type = "Task" [ 822.217077] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.225918] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352445, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.346981] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.361574] env[62914]: DEBUG nova.scheduler.client.report [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.378835] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352444, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486874} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.378835] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 9fab500c-344c-46cd-b060-c8f1aa0f1cba/9fab500c-344c-46cd-b060-c8f1aa0f1cba.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 822.378835] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 822.378835] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64e03c57-1d88-42eb-b342-9bf0b6b154db {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.385182] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Waiting for the task: (returnval){ [ 822.385182] env[62914]: value = "task-1352446" [ 822.385182] env[62914]: _type = "Task" [ 822.385182] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.392588] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352446, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.525489] env[62914]: DEBUG nova.compute.manager [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 822.528054] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.528054] env[62914]: DEBUG nova.compute.manager [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Instance network_info: |[{"id": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "address": "fa:16:3e:7a:f0:fa", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb12a55b6-c3", "ovs_interfaceid": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 822.528625] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:f0:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e31264e2-3e0a-4dfb-ba1f-6389d7d47548', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b12a55b6-c396-44ab-a9ef-b183e97189e3', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.535849] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Creating folder: Project (09d51090ac5d41fc966c7e132f288824). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 822.536120] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-59f2f55e-c5c6-4711-842d-55d637c9ee67 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.545700] env[62914]: DEBUG nova.virt.hardware [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 822.545918] env[62914]: DEBUG nova.virt.hardware [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 822.546085] env[62914]: DEBUG nova.virt.hardware [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 822.546267] env[62914]: DEBUG nova.virt.hardware [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 822.546409] env[62914]: DEBUG nova.virt.hardware [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 822.546552] env[62914]: DEBUG nova.virt.hardware [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 822.546761] env[62914]: DEBUG nova.virt.hardware [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 822.546949] env[62914]: DEBUG nova.virt.hardware [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 822.547131] env[62914]: DEBUG nova.virt.hardware [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 822.547291] env[62914]: DEBUG nova.virt.hardware [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 822.547461] env[62914]: DEBUG nova.virt.hardware [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 822.548236] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0102a40c-61ca-438d-a348-5859364e42ab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.551584] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Created folder: Project (09d51090ac5d41fc966c7e132f288824) in parent group-v288131. [ 822.551761] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Creating folder: Instances. Parent ref: group-v288166. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 822.552258] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-015fa94a-6c50-4a6f-b73c-7ee85b7bbcc1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.556592] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11eaa921-c8a5-4ad2-9099-f89ad247ecd6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.561271] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Created folder: Instances in parent group-v288166. [ 822.561489] env[62914]: DEBUG oslo.service.loopingcall [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.561938] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 822.562144] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d1f56998-9624-43e4-a1c3-218cd27bce55 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.588187] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.588187] env[62914]: value = "task-1352449" [ 822.588187] env[62914]: _type = "Task" [ 822.588187] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.596045] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352449, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.727375] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352445, 'name': PowerOffVM_Task, 'duration_secs': 0.128611} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.727773] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 822.727978] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 822.728914] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a07d42-8104-4724-b1f5-6584dcea64b0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.735268] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 822.735541] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66d86f36-b446-41ba-9913-f08e5d803392 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.751205] env[62914]: DEBUG nova.compute.manager [req-ee8332d8-227e-4f40-882a-0ea48dd37967 req-6a84dab8-3edb-4da9-8278-d929f0fdb813 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Received event network-changed-b12a55b6-c396-44ab-a9ef-b183e97189e3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.751454] env[62914]: DEBUG nova.compute.manager [req-ee8332d8-227e-4f40-882a-0ea48dd37967 req-6a84dab8-3edb-4da9-8278-d929f0fdb813 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Refreshing instance network info cache due to event network-changed-b12a55b6-c396-44ab-a9ef-b183e97189e3. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 822.751819] env[62914]: DEBUG oslo_concurrency.lockutils [req-ee8332d8-227e-4f40-882a-0ea48dd37967 req-6a84dab8-3edb-4da9-8278-d929f0fdb813 service nova] Acquiring lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.752112] env[62914]: DEBUG oslo_concurrency.lockutils [req-ee8332d8-227e-4f40-882a-0ea48dd37967 req-6a84dab8-3edb-4da9-8278-d929f0fdb813 service nova] Acquired lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.752367] env[62914]: DEBUG nova.network.neutron [req-ee8332d8-227e-4f40-882a-0ea48dd37967 req-6a84dab8-3edb-4da9-8278-d929f0fdb813 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Refreshing network info cache for port b12a55b6-c396-44ab-a9ef-b183e97189e3 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 822.757549] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 822.757758] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 822.758150] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Deleting the datastore file [datastore1] 87ef6b06-f699-4de3-8b89-854717074406 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 822.758425] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cddaae15-63ce-4bb7-b80b-b493139cd88e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.764826] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 822.764826] env[62914]: value = "task-1352451" [ 822.764826] env[62914]: _type = "Task" [ 822.764826] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.773085] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352451, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.867269] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.362s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.867813] env[62914]: DEBUG nova.compute.manager [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 822.871055] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.313s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.894087] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352446, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062485} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.894367] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 822.895177] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12779649-05dd-479c-a6a9-fe2f183a8bb0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.919404] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 9fab500c-344c-46cd-b060-c8f1aa0f1cba/9fab500c-344c-46cd-b060-c8f1aa0f1cba.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.920749] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ede7aa81-516e-456e-a7d9-ab39ba0570d6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.940549] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Waiting for the task: (returnval){ [ 822.940549] env[62914]: value = "task-1352452" [ 822.940549] env[62914]: _type = "Task" [ 822.940549] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.948707] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352452, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.076701] env[62914]: DEBUG nova.network.neutron [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Successfully updated port: b82c2ddd-9dd3-4bd8-ad01-68787d570158 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 823.097808] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352449, 'name': CreateVM_Task, 'duration_secs': 0.331627} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.097994] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 823.098717] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.098911] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.099241] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 823.099484] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1062880e-548d-48f1-be06-5bb17a5253a6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.103794] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 823.103794] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]526ea16e-8c99-7e62-9325-19445801c2ce" [ 823.103794] env[62914]: _type = "Task" [ 823.103794] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.111546] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]526ea16e-8c99-7e62-9325-19445801c2ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.274396] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352451, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.262128} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.274751] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 823.274947] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 823.275186] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 823.374675] env[62914]: DEBUG nova.compute.utils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 823.376592] env[62914]: DEBUG nova.compute.manager [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 823.376592] env[62914]: DEBUG nova.network.neutron [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 823.427732] env[62914]: DEBUG nova.policy [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '82a857722ff043c994009b16db6373aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f897e1c819f44362857de4d4398ca613', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 823.454805] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352452, 'name': ReconfigVM_Task, 'duration_secs': 0.2844} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.456797] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 9fab500c-344c-46cd-b060-c8f1aa0f1cba/9fab500c-344c-46cd-b060-c8f1aa0f1cba.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.459101] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7f13e308-1dcd-427b-a4af-567207d7be2e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.463973] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Waiting for the task: (returnval){ [ 823.463973] env[62914]: value = "task-1352453" [ 823.463973] env[62914]: _type = "Task" [ 823.463973] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.471730] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352453, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.556129] env[62914]: DEBUG nova.network.neutron [req-ee8332d8-227e-4f40-882a-0ea48dd37967 req-6a84dab8-3edb-4da9-8278-d929f0fdb813 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Updated VIF entry in instance network info cache for port b12a55b6-c396-44ab-a9ef-b183e97189e3. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 823.556129] env[62914]: DEBUG nova.network.neutron [req-ee8332d8-227e-4f40-882a-0ea48dd37967 req-6a84dab8-3edb-4da9-8278-d929f0fdb813 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Updating instance_info_cache with network_info: [{"id": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "address": "fa:16:3e:7a:f0:fa", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb12a55b6-c3", "ovs_interfaceid": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.582653] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "refresh_cache-9fec3e97-4203-4a90-8a43-600d11d7e7a0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.583272] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "refresh_cache-9fec3e97-4203-4a90-8a43-600d11d7e7a0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.583272] env[62914]: DEBUG nova.network.neutron [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 823.614709] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]526ea16e-8c99-7e62-9325-19445801c2ce, 'name': SearchDatastore_Task, 'duration_secs': 0.008682} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.617417] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.617660] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 823.617887] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.618066] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.618257] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.618699] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-58b9db3f-dbff-4feb-bf25-b89ddeee70c8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.626416] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.626623] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 823.627443] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb75435a-f9e4-4526-9c39-24b51a12fbd9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.634809] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 823.634809] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5253ea1a-1d2d-2f71-68fa-39ec4081fff9" [ 823.634809] env[62914]: _type = "Task" [ 823.634809] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.642956] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5253ea1a-1d2d-2f71-68fa-39ec4081fff9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.644753] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb021faa-b4fe-42aa-90b2-67d7badbbeed {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.650868] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af08acfe-4d45-4593-8b6d-bbbade06432c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.681595] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df1bb58-78f8-4fd5-9738-4ecae0d40582 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.689246] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2eafb61-fa29-4622-8c53-e7c8b8f1766f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.702783] env[62914]: DEBUG nova.compute.provider_tree [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.736918] env[62914]: DEBUG nova.network.neutron [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Successfully created port: 4f7e867d-96a4-4d4e-82fa-43469303566c {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.881985] env[62914]: DEBUG nova.compute.manager [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 823.973623] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352453, 'name': Rename_Task, 'duration_secs': 0.142965} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.973895] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 823.974158] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-622bad48-dbc8-49cf-abb8-fa9bced58c2e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.980389] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Waiting for the task: (returnval){ [ 823.980389] env[62914]: value = "task-1352454" [ 823.980389] env[62914]: _type = "Task" [ 823.980389] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.987652] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352454, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.057956] env[62914]: DEBUG oslo_concurrency.lockutils [req-ee8332d8-227e-4f40-882a-0ea48dd37967 req-6a84dab8-3edb-4da9-8278-d929f0fdb813 service nova] Releasing lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.117226] env[62914]: DEBUG nova.network.neutron [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.144725] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5253ea1a-1d2d-2f71-68fa-39ec4081fff9, 'name': SearchDatastore_Task, 'duration_secs': 0.008528} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.145496] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91272c57-f13a-4145-8298-252017f568ee {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.150162] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 824.150162] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a1e833-65b1-ce4c-61db-57fc2320f59b" [ 824.150162] env[62914]: _type = "Task" [ 824.150162] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.159560] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a1e833-65b1-ce4c-61db-57fc2320f59b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.206160] env[62914]: DEBUG nova.scheduler.client.report [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 824.250791] env[62914]: DEBUG nova.network.neutron [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Updating instance_info_cache with network_info: [{"id": "b82c2ddd-9dd3-4bd8-ad01-68787d570158", "address": "fa:16:3e:5f:a6:b2", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb82c2ddd-9d", "ovs_interfaceid": "b82c2ddd-9dd3-4bd8-ad01-68787d570158", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.310920] env[62914]: DEBUG nova.virt.hardware [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.311190] env[62914]: DEBUG nova.virt.hardware [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.311342] env[62914]: DEBUG nova.virt.hardware [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.311518] env[62914]: DEBUG nova.virt.hardware [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.311660] env[62914]: DEBUG nova.virt.hardware [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.311807] env[62914]: DEBUG nova.virt.hardware [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.312014] env[62914]: DEBUG nova.virt.hardware [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.312175] env[62914]: DEBUG nova.virt.hardware [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.312336] env[62914]: DEBUG nova.virt.hardware [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.312491] env[62914]: DEBUG nova.virt.hardware [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.312654] env[62914]: DEBUG nova.virt.hardware [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.313570] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd38376-6e21-474b-9716-267c0d34171d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.321594] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b061cfaa-3bd3-419a-a8a6-6111e71ccc36 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.335480] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Instance VIF info [] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 824.341185] env[62914]: DEBUG oslo.service.loopingcall [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.341450] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 824.341690] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b626a68-b9ef-4015-b758-3d62b83ca6c4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.358310] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 824.358310] env[62914]: value = "task-1352455" [ 824.358310] env[62914]: _type = "Task" [ 824.358310] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.366139] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352455, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.491162] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352454, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.660765] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a1e833-65b1-ce4c-61db-57fc2320f59b, 'name': SearchDatastore_Task, 'duration_secs': 0.013285} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.661051] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.661350] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07/0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 824.661614] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4a3265b-4c0f-4dfd-8e7b-09532f04c64f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.668169] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 824.668169] env[62914]: value = "task-1352456" [ 824.668169] env[62914]: _type = "Task" [ 824.668169] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.676695] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352456, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.711450] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.840s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.712111] env[62914]: ERROR nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f326bcf5-2bc0-42f3-ae49-60e6029eab73, please check neutron logs for more information. [ 824.712111] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Traceback (most recent call last): [ 824.712111] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 824.712111] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] self.driver.spawn(context, instance, image_meta, [ 824.712111] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 824.712111] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 824.712111] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 824.712111] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] vm_ref = self.build_virtual_machine(instance, [ 824.712111] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 824.712111] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] vif_infos = vmwarevif.get_vif_info(self._session, [ 824.712111] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 824.712448] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] for vif in network_info: [ 824.712448] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 824.712448] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] return self._sync_wrapper(fn, *args, **kwargs) [ 824.712448] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 824.712448] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] self.wait() [ 824.712448] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 824.712448] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] self[:] = self._gt.wait() [ 824.712448] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 824.712448] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] return self._exit_event.wait() [ 824.712448] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 824.712448] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] result = hub.switch() [ 824.712448] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 824.712448] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] return self.greenlet.switch() [ 824.712809] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 824.712809] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] result = function(*args, **kwargs) [ 824.712809] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 824.712809] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] return func(*args, **kwargs) [ 824.712809] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 824.712809] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] raise e [ 824.712809] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 824.712809] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] nwinfo = self.network_api.allocate_for_instance( [ 824.712809] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 824.712809] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] created_port_ids = self._update_ports_for_instance( [ 824.712809] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 824.712809] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] with excutils.save_and_reraise_exception(): [ 824.712809] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 824.713195] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] self.force_reraise() [ 824.713195] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 824.713195] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] raise self.value [ 824.713195] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 824.713195] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] updated_port = self._update_port( [ 824.713195] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 824.713195] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] _ensure_no_port_binding_failure(port) [ 824.713195] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 824.713195] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] raise exception.PortBindingFailed(port_id=port['id']) [ 824.713195] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] nova.exception.PortBindingFailed: Binding failed for port f326bcf5-2bc0-42f3-ae49-60e6029eab73, please check neutron logs for more information. [ 824.713195] env[62914]: ERROR nova.compute.manager [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] [ 824.713510] env[62914]: DEBUG nova.compute.utils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Binding failed for port f326bcf5-2bc0-42f3-ae49-60e6029eab73, please check neutron logs for more information. {{(pid=62914) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 824.714274] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.424s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.715741] env[62914]: INFO nova.compute.claims [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 824.718568] env[62914]: DEBUG nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Build of instance 8a342650-70f0-4d5b-8086-6450dcc8d95b was re-scheduled: Binding failed for port f326bcf5-2bc0-42f3-ae49-60e6029eab73, please check neutron logs for more information. {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 824.718828] env[62914]: DEBUG nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Unplugging VIFs for instance {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 824.719250] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Acquiring lock "refresh_cache-8a342650-70f0-4d5b-8086-6450dcc8d95b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.719250] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Acquired lock "refresh_cache-8a342650-70f0-4d5b-8086-6450dcc8d95b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.719426] env[62914]: DEBUG nova.network.neutron [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 824.753801] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "refresh_cache-9fec3e97-4203-4a90-8a43-600d11d7e7a0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.754147] env[62914]: DEBUG nova.compute.manager [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Instance network_info: |[{"id": "b82c2ddd-9dd3-4bd8-ad01-68787d570158", "address": "fa:16:3e:5f:a6:b2", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb82c2ddd-9d", "ovs_interfaceid": "b82c2ddd-9dd3-4bd8-ad01-68787d570158", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 824.754984] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5f:a6:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a8c8175-1197-4f12-baac-ef6aba95f585', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b82c2ddd-9dd3-4bd8-ad01-68787d570158', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 824.762530] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Creating folder: Project (1dcda44f0d6a455bb17966e7c5061190). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 824.763088] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b72789e-6b95-448c-b93c-29f266b20815 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.773788] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Created folder: Project (1dcda44f0d6a455bb17966e7c5061190) in parent group-v288131. [ 824.773985] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Creating folder: Instances. Parent ref: group-v288170. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 824.774240] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bac0138e-0872-46a4-8804-8f00529becab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.783390] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Created folder: Instances in parent group-v288170. [ 824.783615] env[62914]: DEBUG oslo.service.loopingcall [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.783796] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 824.783989] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-973444f7-20c1-4ce1-b64e-4627d72bc76e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.800298] env[62914]: DEBUG nova.compute.manager [req-03c1a692-9054-4ff6-9739-7998ef0edd46 req-4571dd9c-30db-431c-9eb5-8b536c95e677 service nova] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Received event network-vif-plugged-b82c2ddd-9dd3-4bd8-ad01-68787d570158 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.800498] env[62914]: DEBUG oslo_concurrency.lockutils [req-03c1a692-9054-4ff6-9739-7998ef0edd46 req-4571dd9c-30db-431c-9eb5-8b536c95e677 service nova] Acquiring lock "9fec3e97-4203-4a90-8a43-600d11d7e7a0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.800789] env[62914]: DEBUG oslo_concurrency.lockutils [req-03c1a692-9054-4ff6-9739-7998ef0edd46 req-4571dd9c-30db-431c-9eb5-8b536c95e677 service nova] Lock "9fec3e97-4203-4a90-8a43-600d11d7e7a0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.800851] env[62914]: DEBUG oslo_concurrency.lockutils [req-03c1a692-9054-4ff6-9739-7998ef0edd46 req-4571dd9c-30db-431c-9eb5-8b536c95e677 service nova] Lock "9fec3e97-4203-4a90-8a43-600d11d7e7a0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.801053] env[62914]: DEBUG nova.compute.manager [req-03c1a692-9054-4ff6-9739-7998ef0edd46 req-4571dd9c-30db-431c-9eb5-8b536c95e677 service nova] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] No waiting events found dispatching network-vif-plugged-b82c2ddd-9dd3-4bd8-ad01-68787d570158 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 824.801240] env[62914]: WARNING nova.compute.manager [req-03c1a692-9054-4ff6-9739-7998ef0edd46 req-4571dd9c-30db-431c-9eb5-8b536c95e677 service nova] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Received unexpected event network-vif-plugged-b82c2ddd-9dd3-4bd8-ad01-68787d570158 for instance with vm_state building and task_state spawning. [ 824.801398] env[62914]: DEBUG nova.compute.manager [req-03c1a692-9054-4ff6-9739-7998ef0edd46 req-4571dd9c-30db-431c-9eb5-8b536c95e677 service nova] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Received event network-changed-b82c2ddd-9dd3-4bd8-ad01-68787d570158 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.801549] env[62914]: DEBUG nova.compute.manager [req-03c1a692-9054-4ff6-9739-7998ef0edd46 req-4571dd9c-30db-431c-9eb5-8b536c95e677 service nova] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Refreshing instance network info cache due to event network-changed-b82c2ddd-9dd3-4bd8-ad01-68787d570158. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 824.801724] env[62914]: DEBUG oslo_concurrency.lockutils [req-03c1a692-9054-4ff6-9739-7998ef0edd46 req-4571dd9c-30db-431c-9eb5-8b536c95e677 service nova] Acquiring lock "refresh_cache-9fec3e97-4203-4a90-8a43-600d11d7e7a0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.801857] env[62914]: DEBUG oslo_concurrency.lockutils [req-03c1a692-9054-4ff6-9739-7998ef0edd46 req-4571dd9c-30db-431c-9eb5-8b536c95e677 service nova] Acquired lock "refresh_cache-9fec3e97-4203-4a90-8a43-600d11d7e7a0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.802067] env[62914]: DEBUG nova.network.neutron [req-03c1a692-9054-4ff6-9739-7998ef0edd46 req-4571dd9c-30db-431c-9eb5-8b536c95e677 service nova] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Refreshing network info cache for port b82c2ddd-9dd3-4bd8-ad01-68787d570158 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 824.808308] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 824.808308] env[62914]: value = "task-1352459" [ 824.808308] env[62914]: _type = "Task" [ 824.808308] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.816770] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352459, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.868627] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352455, 'name': CreateVM_Task, 'duration_secs': 0.281427} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.868816] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 824.869281] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.869442] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.869784] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 824.870385] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b20215b4-8cbc-43cf-8f47-326040384a37 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.875620] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 824.875620] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52b88718-58a9-c684-c89e-8761b228f2ba" [ 824.875620] env[62914]: _type = "Task" [ 824.875620] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.883992] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52b88718-58a9-c684-c89e-8761b228f2ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.894385] env[62914]: DEBUG nova.compute.manager [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 824.917090] env[62914]: DEBUG nova.virt.hardware [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.917474] env[62914]: DEBUG nova.virt.hardware [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.917654] env[62914]: DEBUG nova.virt.hardware [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.917839] env[62914]: DEBUG nova.virt.hardware [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.917983] env[62914]: DEBUG nova.virt.hardware [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.918138] env[62914]: DEBUG nova.virt.hardware [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.918346] env[62914]: DEBUG nova.virt.hardware [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.918505] env[62914]: DEBUG nova.virt.hardware [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.918715] env[62914]: DEBUG nova.virt.hardware [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.918910] env[62914]: DEBUG nova.virt.hardware [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.919104] env[62914]: DEBUG nova.virt.hardware [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.919994] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8291e68-4c12-47aa-a019-b035fad86e75 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.927769] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45e43ed-4097-4ff0-9857-52fb2d687264 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.991938] env[62914]: DEBUG oslo_vmware.api [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352454, 'name': PowerOnVM_Task, 'duration_secs': 0.575352} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.992244] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 824.992461] env[62914]: INFO nova.compute.manager [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Took 7.21 seconds to spawn the instance on the hypervisor. [ 824.992640] env[62914]: DEBUG nova.compute.manager [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 824.993450] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120cd747-a2ec-4c78-8861-06abfef4cda6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.179015] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352456, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.220377] env[62914]: DEBUG nova.compute.manager [req-b38361c0-2b70-472c-9edc-67036dbbe2c4 req-8c2900cf-60d3-4b28-a959-b778c7095225 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Received event network-vif-plugged-4f7e867d-96a4-4d4e-82fa-43469303566c {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.220377] env[62914]: DEBUG oslo_concurrency.lockutils [req-b38361c0-2b70-472c-9edc-67036dbbe2c4 req-8c2900cf-60d3-4b28-a959-b778c7095225 service nova] Acquiring lock "46d139a1-a4ae-435e-9d6f-cfc06d706128-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.220377] env[62914]: DEBUG oslo_concurrency.lockutils [req-b38361c0-2b70-472c-9edc-67036dbbe2c4 req-8c2900cf-60d3-4b28-a959-b778c7095225 service nova] Lock "46d139a1-a4ae-435e-9d6f-cfc06d706128-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.221031] env[62914]: DEBUG oslo_concurrency.lockutils [req-b38361c0-2b70-472c-9edc-67036dbbe2c4 req-8c2900cf-60d3-4b28-a959-b778c7095225 service nova] Lock "46d139a1-a4ae-435e-9d6f-cfc06d706128-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.221031] env[62914]: DEBUG nova.compute.manager [req-b38361c0-2b70-472c-9edc-67036dbbe2c4 req-8c2900cf-60d3-4b28-a959-b778c7095225 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] No waiting events found dispatching network-vif-plugged-4f7e867d-96a4-4d4e-82fa-43469303566c {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 825.221189] env[62914]: WARNING nova.compute.manager [req-b38361c0-2b70-472c-9edc-67036dbbe2c4 req-8c2900cf-60d3-4b28-a959-b778c7095225 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Received unexpected event network-vif-plugged-4f7e867d-96a4-4d4e-82fa-43469303566c for instance with vm_state building and task_state spawning. [ 825.243029] env[62914]: DEBUG nova.network.neutron [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.303865] env[62914]: DEBUG nova.network.neutron [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Successfully updated port: 4f7e867d-96a4-4d4e-82fa-43469303566c {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 825.322476] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352459, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.337673] env[62914]: DEBUG nova.network.neutron [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.385163] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52b88718-58a9-c684-c89e-8761b228f2ba, 'name': SearchDatastore_Task, 'duration_secs': 0.015409} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.385495] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.385719] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 825.386019] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.386181] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.386360] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 825.386611] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-780fc8f4-9f7d-4ee9-b3d6-07013de07a8e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.401114] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 825.401300] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 825.402017] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d98aec5-6011-4dc4-acba-efb0ecf36c69 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.407794] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 825.407794] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5259ddd9-fe41-7734-0ddb-51669950af40" [ 825.407794] env[62914]: _type = "Task" [ 825.407794] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.418613] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5259ddd9-fe41-7734-0ddb-51669950af40, 'name': SearchDatastore_Task, 'duration_secs': 0.008132} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.419429] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3a1272e-130e-4de7-a5c7-56ea689dd5bc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.423854] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 825.423854] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]528cfaf5-24b6-243e-bc57-81f0dd9b6ecf" [ 825.423854] env[62914]: _type = "Task" [ 825.423854] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.431082] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]528cfaf5-24b6-243e-bc57-81f0dd9b6ecf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.511719] env[62914]: INFO nova.compute.manager [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Took 24.96 seconds to build instance. [ 825.514342] env[62914]: DEBUG nova.network.neutron [req-03c1a692-9054-4ff6-9739-7998ef0edd46 req-4571dd9c-30db-431c-9eb5-8b536c95e677 service nova] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Updated VIF entry in instance network info cache for port b82c2ddd-9dd3-4bd8-ad01-68787d570158. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 825.514657] env[62914]: DEBUG nova.network.neutron [req-03c1a692-9054-4ff6-9739-7998ef0edd46 req-4571dd9c-30db-431c-9eb5-8b536c95e677 service nova] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Updating instance_info_cache with network_info: [{"id": "b82c2ddd-9dd3-4bd8-ad01-68787d570158", "address": "fa:16:3e:5f:a6:b2", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb82c2ddd-9d", "ovs_interfaceid": "b82c2ddd-9dd3-4bd8-ad01-68787d570158", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.678345] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352456, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542402} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.679031] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07/0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 825.679031] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 825.679181] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03d5f7e0-6787-4fbf-937e-d94a9407d0c6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.688509] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 825.688509] env[62914]: value = "task-1352460" [ 825.688509] env[62914]: _type = "Task" [ 825.688509] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.697364] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352460, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.810536] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.810536] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquired lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.810692] env[62914]: DEBUG nova.network.neutron [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 825.823521] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352459, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.842021] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Releasing lock "refresh_cache-8a342650-70f0-4d5b-8086-6450dcc8d95b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.842021] env[62914]: DEBUG nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62914) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 825.842021] env[62914]: DEBUG nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 825.842021] env[62914]: DEBUG nova.network.neutron [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 825.857612] env[62914]: DEBUG nova.network.neutron [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.933765] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]528cfaf5-24b6-243e-bc57-81f0dd9b6ecf, 'name': SearchDatastore_Task, 'duration_secs': 0.008663} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.935992] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.936269] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 87ef6b06-f699-4de3-8b89-854717074406/87ef6b06-f699-4de3-8b89-854717074406.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 825.936678] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-047bf559-1387-4b97-a55d-c26c3c461acc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.943154] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 825.943154] env[62914]: value = "task-1352461" [ 825.943154] env[62914]: _type = "Task" [ 825.943154] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.954933] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352461, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.959592] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b045421-fb8f-46b5-8aaf-960089aca313 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.966462] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b0541c-a2c9-4ca8-957d-95d42e5339f0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.995668] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611ab55f-81e6-4479-9a17-8ccf42a3dbc7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.003554] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fb8d25b-4f94-44ba-99ac-0116a0a98fc2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.016495] env[62914]: DEBUG oslo_concurrency.lockutils [None req-93561105-b405-49dc-8ec8-114c7d5dbdf5 tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Lock "9fab500c-344c-46cd-b060-c8f1aa0f1cba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.950s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.016922] env[62914]: DEBUG oslo_concurrency.lockutils [req-03c1a692-9054-4ff6-9739-7998ef0edd46 req-4571dd9c-30db-431c-9eb5-8b536c95e677 service nova] Releasing lock "refresh_cache-9fec3e97-4203-4a90-8a43-600d11d7e7a0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.017481] env[62914]: DEBUG nova.compute.provider_tree [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.199191] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352460, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064075} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.199455] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 826.200358] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3202e9af-5d81-4d1d-9c86-133c5657b4d7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.222273] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07/0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.222565] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41570fa8-5ca3-4ee2-9f44-7f1cb3470adb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.243516] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 826.243516] env[62914]: value = "task-1352462" [ 826.243516] env[62914]: _type = "Task" [ 826.243516] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.252396] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352462, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.325374] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352459, 'name': CreateVM_Task, 'duration_secs': 1.112472} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.325566] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 826.326528] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.326691] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.327062] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 826.327322] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af986f19-f811-4adc-a2eb-8ececf229c47 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.332229] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 826.332229] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]524d99ed-82c9-d974-ba11-5c1670972462" [ 826.332229] env[62914]: _type = "Task" [ 826.332229] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.340320] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]524d99ed-82c9-d974-ba11-5c1670972462, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.360048] env[62914]: DEBUG nova.network.neutron [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.399641] env[62914]: DEBUG nova.network.neutron [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.453821] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352461, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.520630] env[62914]: DEBUG nova.scheduler.client.report [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 826.523827] env[62914]: DEBUG nova.compute.manager [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 826.755246] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352462, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.849041] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]524d99ed-82c9-d974-ba11-5c1670972462, 'name': SearchDatastore_Task, 'duration_secs': 0.018056} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.849760] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.849760] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.849961] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.850413] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.850413] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.850615] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e929338-748c-4d70-bcd1-debf491d60ed {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.865110] env[62914]: INFO nova.compute.manager [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] [instance: 8a342650-70f0-4d5b-8086-6450dcc8d95b] Took 1.02 seconds to deallocate network for instance. [ 826.872197] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.872525] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 826.873376] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4525fe20-a841-49ec-9fd3-a93502a760b3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.879507] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 826.879507] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d8ab87-0f09-d2d2-3b24-b8c464d31e91" [ 826.879507] env[62914]: _type = "Task" [ 826.879507] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.889493] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d8ab87-0f09-d2d2-3b24-b8c464d31e91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.935293] env[62914]: DEBUG nova.network.neutron [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Updating instance_info_cache with network_info: [{"id": "4f7e867d-96a4-4d4e-82fa-43469303566c", "address": "fa:16:3e:51:58:7b", "network": {"id": "c8cdb39e-9415-4f67-8199-fa1b043d04d8", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-373473409-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f897e1c819f44362857de4d4398ca613", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f7e867d-96", "ovs_interfaceid": "4f7e867d-96a4-4d4e-82fa-43469303566c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.954393] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352461, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.995209} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.954709] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 87ef6b06-f699-4de3-8b89-854717074406/87ef6b06-f699-4de3-8b89-854717074406.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 826.954983] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 826.955296] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-44ff9c66-2b34-421b-afea-fa78a1f278e9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.963791] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 826.963791] env[62914]: value = "task-1352463" [ 826.963791] env[62914]: _type = "Task" [ 826.963791] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.973814] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352463, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.027640] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.313s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.031461] env[62914]: DEBUG nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 827.034381] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.888s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.036050] env[62914]: INFO nova.compute.claims [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.063816] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.252769] env[62914]: DEBUG nova.compute.manager [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Received event network-changed-4f7e867d-96a4-4d4e-82fa-43469303566c {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.252897] env[62914]: DEBUG nova.compute.manager [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Refreshing instance network info cache due to event network-changed-4f7e867d-96a4-4d4e-82fa-43469303566c. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 827.253123] env[62914]: DEBUG oslo_concurrency.lockutils [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] Acquiring lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.256542] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352462, 'name': ReconfigVM_Task, 'duration_secs': 0.757826} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.256776] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07/0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.257374] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a8f73b2d-83a6-4293-891e-8a7e0d08253c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.264809] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 827.264809] env[62914]: value = "task-1352464" [ 827.264809] env[62914]: _type = "Task" [ 827.264809] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.273609] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352464, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.392533] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d8ab87-0f09-d2d2-3b24-b8c464d31e91, 'name': SearchDatastore_Task, 'duration_secs': 0.02847} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.393353] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c4b003e-afc2-4d3a-8ad1-5888384d36d6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.398688] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 827.398688] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]520eb6af-4bf7-a62c-7e24-b7a9a4c67632" [ 827.398688] env[62914]: _type = "Task" [ 827.398688] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.406673] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]520eb6af-4bf7-a62c-7e24-b7a9a4c67632, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.439922] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Releasing lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.440277] env[62914]: DEBUG nova.compute.manager [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Instance network_info: |[{"id": "4f7e867d-96a4-4d4e-82fa-43469303566c", "address": "fa:16:3e:51:58:7b", "network": {"id": "c8cdb39e-9415-4f67-8199-fa1b043d04d8", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-373473409-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f897e1c819f44362857de4d4398ca613", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f7e867d-96", "ovs_interfaceid": "4f7e867d-96a4-4d4e-82fa-43469303566c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 827.440568] env[62914]: DEBUG oslo_concurrency.lockutils [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] Acquired lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.440742] env[62914]: DEBUG nova.network.neutron [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Refreshing network info cache for port 4f7e867d-96a4-4d4e-82fa-43469303566c {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 827.442260] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:58:7b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '09bf081b-cdf0-4977-abe2-2339a87409ab', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4f7e867d-96a4-4d4e-82fa-43469303566c', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 827.451017] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Creating folder: Project (f897e1c819f44362857de4d4398ca613). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 827.453992] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d8d6d7a8-4658-4147-9875-157329b55246 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.465077] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Created folder: Project (f897e1c819f44362857de4d4398ca613) in parent group-v288131. [ 827.465277] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Creating folder: Instances. Parent ref: group-v288173. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 827.468395] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0cfb6010-499a-485f-b568-a6c806d9d132 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.474627] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352463, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066998} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.474872] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 827.475662] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-948fe5b6-3c42-4269-b276-821cae8e0226 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.479447] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Created folder: Instances in parent group-v288173. [ 827.479447] env[62914]: DEBUG oslo.service.loopingcall [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.479914] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 827.480197] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4b1bf407-82da-4484-8d0b-3c43551e25f6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.512268] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 87ef6b06-f699-4de3-8b89-854717074406/87ef6b06-f699-4de3-8b89-854717074406.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 827.513013] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40276207-10ea-484d-82c1-fbc397e16d6d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.529545] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 827.529545] env[62914]: value = "task-1352467" [ 827.529545] env[62914]: _type = "Task" [ 827.529545] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.534361] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 827.534361] env[62914]: value = "task-1352468" [ 827.534361] env[62914]: _type = "Task" [ 827.534361] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.542319] env[62914]: DEBUG nova.compute.utils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.545739] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352467, 'name': CreateVM_Task} progress is 10%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.546286] env[62914]: DEBUG nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 827.546468] env[62914]: DEBUG nova.network.neutron [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 827.552116] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352468, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.594580] env[62914]: DEBUG nova.policy [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd9f93f76e6db40a3924a5429ab2a23f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b6b4c54c48514783a044a75dcdea02b8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 827.740038] env[62914]: DEBUG nova.network.neutron [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Updated VIF entry in instance network info cache for port 4f7e867d-96a4-4d4e-82fa-43469303566c. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 827.740038] env[62914]: DEBUG nova.network.neutron [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Updating instance_info_cache with network_info: [{"id": "4f7e867d-96a4-4d4e-82fa-43469303566c", "address": "fa:16:3e:51:58:7b", "network": {"id": "c8cdb39e-9415-4f67-8199-fa1b043d04d8", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-373473409-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f897e1c819f44362857de4d4398ca613", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f7e867d-96", "ovs_interfaceid": "4f7e867d-96a4-4d4e-82fa-43469303566c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.775244] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352464, 'name': Rename_Task, 'duration_secs': 0.256182} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.775482] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 827.777444] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a0874a5-c231-449e-aa16-883aea8aeff3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.784034] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 827.784034] env[62914]: value = "task-1352469" [ 827.784034] env[62914]: _type = "Task" [ 827.784034] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.795125] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352469, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.915166] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]520eb6af-4bf7-a62c-7e24-b7a9a4c67632, 'name': SearchDatastore_Task, 'duration_secs': 0.010905} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.915166] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.915166] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 9fec3e97-4203-4a90-8a43-600d11d7e7a0/9fec3e97-4203-4a90-8a43-600d11d7e7a0.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 827.915166] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36ed0eba-e250-433f-8e64-933f2dae96b3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.916790] env[62914]: DEBUG nova.network.neutron [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Successfully created port: 1c71fa9c-a921-4649-81d3-9de5b41619c6 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.920845] env[62914]: INFO nova.scheduler.client.report [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Deleted allocations for instance 8a342650-70f0-4d5b-8086-6450dcc8d95b [ 827.935643] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 827.935643] env[62914]: value = "task-1352470" [ 827.935643] env[62914]: _type = "Task" [ 827.935643] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.946396] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352470, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.043889] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352467, 'name': CreateVM_Task, 'duration_secs': 0.326308} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.045477] env[62914]: DEBUG nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 828.049961] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 828.054596] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.054848] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.055245] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.059038] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352468, 'name': ReconfigVM_Task, 'duration_secs': 0.358929} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.059038] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3495db29-a5c8-482f-bd20-0ede1a4679e8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.061145] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 87ef6b06-f699-4de3-8b89-854717074406/87ef6b06-f699-4de3-8b89-854717074406.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 828.061917] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cf956b7a-7e14-42d1-bdf3-8ab1ea0f43c5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.066657] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Waiting for the task: (returnval){ [ 828.066657] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]525522fc-ece0-ae45-861c-39aefe2c2f92" [ 828.066657] env[62914]: _type = "Task" [ 828.066657] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.068118] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 828.068118] env[62914]: value = "task-1352471" [ 828.068118] env[62914]: _type = "Task" [ 828.068118] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.080640] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]525522fc-ece0-ae45-861c-39aefe2c2f92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.085216] env[62914]: DEBUG nova.scheduler.client.report [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Refreshing inventories for resource provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 828.087436] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352471, 'name': Rename_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.102464] env[62914]: DEBUG nova.scheduler.client.report [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Updating ProviderTree inventory for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 828.102613] env[62914]: DEBUG nova.compute.provider_tree [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Updating inventory in ProviderTree for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 828.116855] env[62914]: DEBUG nova.scheduler.client.report [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Refreshing aggregate associations for resource provider 8e1b3185-5ae3-4780-aee8-c507b157ad13, aggregates: None {{(pid=62914) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 828.137931] env[62914]: DEBUG nova.scheduler.client.report [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Refreshing trait associations for resource provider 8e1b3185-5ae3-4780-aee8-c507b157ad13, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62914) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 828.244773] env[62914]: DEBUG oslo_concurrency.lockutils [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] Releasing lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.244773] env[62914]: DEBUG nova.compute.manager [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Received event network-changed-299c1132-e4a5-4803-8e8d-1478b5928463 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.244773] env[62914]: DEBUG nova.compute.manager [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Refreshing instance network info cache due to event network-changed-299c1132-e4a5-4803-8e8d-1478b5928463. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 828.244773] env[62914]: DEBUG oslo_concurrency.lockutils [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] Acquiring lock "refresh_cache-9fab500c-344c-46cd-b060-c8f1aa0f1cba" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.244773] env[62914]: DEBUG oslo_concurrency.lockutils [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] Acquired lock "refresh_cache-9fab500c-344c-46cd-b060-c8f1aa0f1cba" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.245123] env[62914]: DEBUG nova.network.neutron [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Refreshing network info cache for port 299c1132-e4a5-4803-8e8d-1478b5928463 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 828.296432] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352469, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.346956] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c272048-b662-4894-8510-02e02b2924ad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.354203] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bfabc9d-154e-4ad8-a41c-b31b3b02a47b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.382619] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af4c30e-7846-4465-9c91-681e90218047 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.389097] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c33f70-7c2b-4a10-8954-c25d6475375c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.403431] env[62914]: DEBUG nova.compute.provider_tree [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.432420] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f013821e-0498-4030-8acf-dd9234adeb60 tempest-ServersTestMultiNic-553169784 tempest-ServersTestMultiNic-553169784-project-member] Lock "8a342650-70f0-4d5b-8086-6450dcc8d95b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.169s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.444547] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352470, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.580438] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]525522fc-ece0-ae45-861c-39aefe2c2f92, 'name': SearchDatastore_Task, 'duration_secs': 0.017001} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.583375] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.583604] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.583834] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.583976] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.584173] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.584437] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352471, 'name': Rename_Task, 'duration_secs': 0.138327} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.584631] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a71d5ede-c45f-48ca-b313-9f25861b887c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.586843] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 828.586843] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b35f2892-6d47-4ff5-87a7-bda4d0703b31 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.593786] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 828.593786] env[62914]: value = "task-1352472" [ 828.593786] env[62914]: _type = "Task" [ 828.593786] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.599718] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 828.599718] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 828.599949] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b945f4ca-d757-4125-9389-192080e9a2d9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.604789] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352472, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.607776] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Waiting for the task: (returnval){ [ 828.607776] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d60aac-4e02-cd80-f84a-7793ad0ba727" [ 828.607776] env[62914]: _type = "Task" [ 828.607776] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.616295] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d60aac-4e02-cd80-f84a-7793ad0ba727, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.794815] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352469, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.907011] env[62914]: DEBUG nova.scheduler.client.report [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.934835] env[62914]: DEBUG nova.compute.manager [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 828.947628] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352470, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.071680] env[62914]: DEBUG nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 829.076353] env[62914]: DEBUG nova.network.neutron [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Updated VIF entry in instance network info cache for port 299c1132-e4a5-4803-8e8d-1478b5928463. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 829.076718] env[62914]: DEBUG nova.network.neutron [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Updating instance_info_cache with network_info: [{"id": "299c1132-e4a5-4803-8e8d-1478b5928463", "address": "fa:16:3e:d2:24:e5", "network": {"id": "3f1a31ed-b39a-473d-a7f4-979cde7098b0", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-677647149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb17042d6ab4ab4824e39bf809960da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0dd3c126-9d86-4f9a-b81c-e9627c7a5401", "external-id": "nsx-vlan-transportzone-24", "segmentation_id": 24, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap299c1132-e4", "ovs_interfaceid": "299c1132-e4a5-4803-8e8d-1478b5928463", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.109298] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352472, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.111484] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 829.111704] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 829.111856] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.112049] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 829.112234] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.112348] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 829.112536] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 829.112682] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 829.112839] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 829.113023] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 829.113180] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.117059] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9a5708-61bb-4a03-940e-0fb809580cdf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.128117] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14f1032-e2af-4c45-98d3-5604e4cbfa17 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.133798] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d60aac-4e02-cd80-f84a-7793ad0ba727, 'name': SearchDatastore_Task, 'duration_secs': 0.009761} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.135258] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b44d826-a1d5-4b08-8e1f-b5f78e7aaa62 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.149666] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Waiting for the task: (returnval){ [ 829.149666] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52409041-8f14-aaa7-9d40-69cb72a6589b" [ 829.149666] env[62914]: _type = "Task" [ 829.149666] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.156771] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52409041-8f14-aaa7-9d40-69cb72a6589b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.297156] env[62914]: DEBUG oslo_vmware.api [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352469, 'name': PowerOnVM_Task, 'duration_secs': 1.422581} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.297156] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 829.297156] env[62914]: INFO nova.compute.manager [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Took 9.14 seconds to spawn the instance on the hypervisor. [ 829.297156] env[62914]: DEBUG nova.compute.manager [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.297156] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b396feca-4be1-4d2e-9d40-6720730976a2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.415401] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.416728] env[62914]: DEBUG nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 829.421199] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.466s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.422623] env[62914]: INFO nova.compute.claims [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.459817] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352470, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.077665} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.459817] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 9fec3e97-4203-4a90-8a43-600d11d7e7a0/9fec3e97-4203-4a90-8a43-600d11d7e7a0.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 829.459817] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 829.459817] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a95726f3-d3c5-4c37-ad84-457afb4b7a58 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.468034] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 829.468034] env[62914]: value = "task-1352473" [ 829.468034] env[62914]: _type = "Task" [ 829.468034] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.477327] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352473, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.479628] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.511570] env[62914]: DEBUG nova.compute.manager [req-8f54fb0e-3bf7-4f33-8f65-12080813d7dd req-fe3ff72e-27c8-4327-a4dc-3de078f07236 service nova] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Received event network-vif-plugged-1c71fa9c-a921-4649-81d3-9de5b41619c6 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.512355] env[62914]: DEBUG oslo_concurrency.lockutils [req-8f54fb0e-3bf7-4f33-8f65-12080813d7dd req-fe3ff72e-27c8-4327-a4dc-3de078f07236 service nova] Acquiring lock "cb89c84f-414b-4dc7-9db5-12cbc30b52b4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.512355] env[62914]: DEBUG oslo_concurrency.lockutils [req-8f54fb0e-3bf7-4f33-8f65-12080813d7dd req-fe3ff72e-27c8-4327-a4dc-3de078f07236 service nova] Lock "cb89c84f-414b-4dc7-9db5-12cbc30b52b4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.512355] env[62914]: DEBUG oslo_concurrency.lockutils [req-8f54fb0e-3bf7-4f33-8f65-12080813d7dd req-fe3ff72e-27c8-4327-a4dc-3de078f07236 service nova] Lock "cb89c84f-414b-4dc7-9db5-12cbc30b52b4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.512355] env[62914]: DEBUG nova.compute.manager [req-8f54fb0e-3bf7-4f33-8f65-12080813d7dd req-fe3ff72e-27c8-4327-a4dc-3de078f07236 service nova] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] No waiting events found dispatching network-vif-plugged-1c71fa9c-a921-4649-81d3-9de5b41619c6 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 829.512756] env[62914]: WARNING nova.compute.manager [req-8f54fb0e-3bf7-4f33-8f65-12080813d7dd req-fe3ff72e-27c8-4327-a4dc-3de078f07236 service nova] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Received unexpected event network-vif-plugged-1c71fa9c-a921-4649-81d3-9de5b41619c6 for instance with vm_state building and task_state spawning. [ 829.579738] env[62914]: DEBUG oslo_concurrency.lockutils [req-fccc50fb-60cd-4c0b-882d-23bd6affe9ea req-4f4d7d7a-35db-449e-8abc-91a5c786fd38 service nova] Releasing lock "refresh_cache-9fab500c-344c-46cd-b060-c8f1aa0f1cba" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.605045] env[62914]: DEBUG oslo_vmware.api [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352472, 'name': PowerOnVM_Task, 'duration_secs': 0.583517} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.605372] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 829.605599] env[62914]: DEBUG nova.compute.manager [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.606441] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6f2476-a257-454e-a47f-62ea413acc50 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.661962] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52409041-8f14-aaa7-9d40-69cb72a6589b, 'name': SearchDatastore_Task, 'duration_secs': 0.02506} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.663339] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.663864] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 46d139a1-a4ae-435e-9d6f-cfc06d706128/46d139a1-a4ae-435e-9d6f-cfc06d706128.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 829.664651] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f5ba8701-48ff-4035-a72e-d6349c43d64a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.672794] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Waiting for the task: (returnval){ [ 829.672794] env[62914]: value = "task-1352474" [ 829.672794] env[62914]: _type = "Task" [ 829.672794] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.683385] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352474, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.815842] env[62914]: INFO nova.compute.manager [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Took 27.45 seconds to build instance. [ 829.934798] env[62914]: DEBUG nova.compute.utils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.934798] env[62914]: DEBUG nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 829.934798] env[62914]: DEBUG nova.network.neutron [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 829.979488] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352473, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.354111} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.983039] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.983039] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc774264-881d-4b7b-850b-bccafe412e08 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.003446] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 9fec3e97-4203-4a90-8a43-600d11d7e7a0/9fec3e97-4203-4a90-8a43-600d11d7e7a0.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 830.005298] env[62914]: DEBUG nova.policy [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd9f93f76e6db40a3924a5429ab2a23f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b6b4c54c48514783a044a75dcdea02b8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 830.006891] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-735b6359-52ee-49bc-b580-647d5e65319a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.025477] env[62914]: DEBUG nova.network.neutron [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Successfully updated port: 1c71fa9c-a921-4649-81d3-9de5b41619c6 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 830.034020] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 830.034020] env[62914]: value = "task-1352475" [ 830.034020] env[62914]: _type = "Task" [ 830.034020] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.042314] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352475, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.044660] env[62914]: DEBUG nova.compute.manager [req-a53500a6-193f-41a6-9904-5e6860b05c59 req-4fc9193b-d607-452c-a003-3910b39743e0 service nova] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Received event network-changed-1c71fa9c-a921-4649-81d3-9de5b41619c6 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.044962] env[62914]: DEBUG nova.compute.manager [req-a53500a6-193f-41a6-9904-5e6860b05c59 req-4fc9193b-d607-452c-a003-3910b39743e0 service nova] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Refreshing instance network info cache due to event network-changed-1c71fa9c-a921-4649-81d3-9de5b41619c6. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 830.045311] env[62914]: DEBUG oslo_concurrency.lockutils [req-a53500a6-193f-41a6-9904-5e6860b05c59 req-4fc9193b-d607-452c-a003-3910b39743e0 service nova] Acquiring lock "refresh_cache-cb89c84f-414b-4dc7-9db5-12cbc30b52b4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.047054] env[62914]: DEBUG oslo_concurrency.lockutils [req-a53500a6-193f-41a6-9904-5e6860b05c59 req-4fc9193b-d607-452c-a003-3910b39743e0 service nova] Acquired lock "refresh_cache-cb89c84f-414b-4dc7-9db5-12cbc30b52b4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.047054] env[62914]: DEBUG nova.network.neutron [req-a53500a6-193f-41a6-9904-5e6860b05c59 req-4fc9193b-d607-452c-a003-3910b39743e0 service nova] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Refreshing network info cache for port 1c71fa9c-a921-4649-81d3-9de5b41619c6 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 830.122379] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.184774] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352474, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.317479] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3bc82dfd-286f-424f-8298-30bfee0aa9d3 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.573s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.439112] env[62914]: DEBUG nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 830.528676] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "refresh_cache-cb89c84f-414b-4dc7-9db5-12cbc30b52b4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.544456] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352475, 'name': ReconfigVM_Task, 'duration_secs': 0.440763} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.545246] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 9fec3e97-4203-4a90-8a43-600d11d7e7a0/9fec3e97-4203-4a90-8a43-600d11d7e7a0.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 830.546286] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d3848f5-6337-4705-96df-f36e1941001d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.560185] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 830.560185] env[62914]: value = "task-1352476" [ 830.560185] env[62914]: _type = "Task" [ 830.560185] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.570077] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352476, 'name': Rename_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.601254] env[62914]: DEBUG nova.network.neutron [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Successfully created port: 68eeb848-a0dd-455b-8501-a7b14519e427 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 830.615744] env[62914]: DEBUG nova.network.neutron [req-a53500a6-193f-41a6-9904-5e6860b05c59 req-4fc9193b-d607-452c-a003-3910b39743e0 service nova] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.690306] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352474, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.994659} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.694119] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 46d139a1-a4ae-435e-9d6f-cfc06d706128/46d139a1-a4ae-435e-9d6f-cfc06d706128.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 830.695699] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 830.695863] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-465be8be-0d84-42be-848a-1aafd41194bc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.705164] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Waiting for the task: (returnval){ [ 830.705164] env[62914]: value = "task-1352477" [ 830.705164] env[62914]: _type = "Task" [ 830.705164] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.714226] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352477, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.729687] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d6d65b-b351-497a-a015-032147883b85 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.737047] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6392fbd9-e999-4f40-932e-8746e82b4aa6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.772510] env[62914]: DEBUG nova.network.neutron [req-a53500a6-193f-41a6-9904-5e6860b05c59 req-4fc9193b-d607-452c-a003-3910b39743e0 service nova] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.774573] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4c11af-bb42-49c8-8dab-8d19a5cc7bac {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.783016] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c0c9d8-8b3e-42bd-8d05-b8e1295f6060 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.798292] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquiring lock "87ef6b06-f699-4de3-8b89-854717074406" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.798755] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Lock "87ef6b06-f699-4de3-8b89-854717074406" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.799117] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquiring lock "87ef6b06-f699-4de3-8b89-854717074406-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.799423] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Lock "87ef6b06-f699-4de3-8b89-854717074406-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.799707] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Lock "87ef6b06-f699-4de3-8b89-854717074406-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.801524] env[62914]: DEBUG nova.compute.provider_tree [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.802975] env[62914]: INFO nova.compute.manager [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Terminating instance [ 830.804766] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquiring lock "refresh_cache-87ef6b06-f699-4de3-8b89-854717074406" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.805060] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquired lock "refresh_cache-87ef6b06-f699-4de3-8b89-854717074406" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.805329] env[62914]: DEBUG nova.network.neutron [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 830.820073] env[62914]: DEBUG nova.compute.manager [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 831.068481] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352476, 'name': Rename_Task, 'duration_secs': 0.196071} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.068779] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 831.068996] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c0d58b2-2086-4b83-bd8d-aebbc9da282d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.075643] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 831.075643] env[62914]: value = "task-1352478" [ 831.075643] env[62914]: _type = "Task" [ 831.075643] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.084771] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352478, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.212846] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352477, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090967} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.213513] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 831.214372] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8347e7b5-8cb3-4d6f-b0b0-f5912e80f7dc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.241350] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 46d139a1-a4ae-435e-9d6f-cfc06d706128/46d139a1-a4ae-435e-9d6f-cfc06d706128.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 831.242058] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ee511a8-03b2-489c-8a11-c559621de9f9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.264430] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Waiting for the task: (returnval){ [ 831.264430] env[62914]: value = "task-1352479" [ 831.264430] env[62914]: _type = "Task" [ 831.264430] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.271469] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352479, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.278304] env[62914]: DEBUG oslo_concurrency.lockutils [req-a53500a6-193f-41a6-9904-5e6860b05c59 req-4fc9193b-d607-452c-a003-3910b39743e0 service nova] Releasing lock "refresh_cache-cb89c84f-414b-4dc7-9db5-12cbc30b52b4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.278468] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquired lock "refresh_cache-cb89c84f-414b-4dc7-9db5-12cbc30b52b4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.278620] env[62914]: DEBUG nova.network.neutron [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.305563] env[62914]: DEBUG nova.scheduler.client.report [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.326357] env[62914]: DEBUG nova.network.neutron [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.340945] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.435304] env[62914]: DEBUG nova.network.neutron [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.445304] env[62914]: DEBUG nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 831.480122] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.480122] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.480493] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.480796] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.481069] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.481499] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.481812] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.482082] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.482449] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.483022] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.483349] env[62914]: DEBUG nova.virt.hardware [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.484748] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-486f6a60-e512-44fe-8af8-42b8a8b042fc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.499359] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048bce5a-3625-413d-9807-7a40ded51feb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.587060] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352478, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.775364] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352479, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.812778] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.813382] env[62914]: DEBUG nova.compute.manager [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 831.816965] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.974s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.818882] env[62914]: INFO nova.compute.claims [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 831.852207] env[62914]: DEBUG nova.network.neutron [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.940359] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Releasing lock "refresh_cache-87ef6b06-f699-4de3-8b89-854717074406" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.940783] env[62914]: DEBUG nova.compute.manager [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 831.940982] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 831.942055] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d081455-da54-4903-b6b2-84bfd726cf73 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.951766] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 831.952426] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dcfc3aea-d62b-4415-81c2-3bfcc16d2518 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.960154] env[62914]: DEBUG oslo_vmware.api [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 831.960154] env[62914]: value = "task-1352480" [ 831.960154] env[62914]: _type = "Task" [ 831.960154] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.971427] env[62914]: DEBUG oslo_vmware.api [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352480, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.032861] env[62914]: DEBUG nova.network.neutron [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Updating instance_info_cache with network_info: [{"id": "1c71fa9c-a921-4649-81d3-9de5b41619c6", "address": "fa:16:3e:db:9a:a4", "network": {"id": "b3def210-2cac-4154-83ca-95b77b5e945c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1178445589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b6b4c54c48514783a044a75dcdea02b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c71fa9c-a9", "ovs_interfaceid": "1c71fa9c-a921-4649-81d3-9de5b41619c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.068808] env[62914]: DEBUG nova.compute.manager [req-bb8f0a1a-d0cf-4465-a8a3-c52f3f5429c8 req-89c005d4-88f7-4712-999c-ccb7f98f6aad service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Received event network-changed-b12a55b6-c396-44ab-a9ef-b183e97189e3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.069178] env[62914]: DEBUG nova.compute.manager [req-bb8f0a1a-d0cf-4465-a8a3-c52f3f5429c8 req-89c005d4-88f7-4712-999c-ccb7f98f6aad service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Refreshing instance network info cache due to event network-changed-b12a55b6-c396-44ab-a9ef-b183e97189e3. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 832.069284] env[62914]: DEBUG oslo_concurrency.lockutils [req-bb8f0a1a-d0cf-4465-a8a3-c52f3f5429c8 req-89c005d4-88f7-4712-999c-ccb7f98f6aad service nova] Acquiring lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.069426] env[62914]: DEBUG oslo_concurrency.lockutils [req-bb8f0a1a-d0cf-4465-a8a3-c52f3f5429c8 req-89c005d4-88f7-4712-999c-ccb7f98f6aad service nova] Acquired lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.069582] env[62914]: DEBUG nova.network.neutron [req-bb8f0a1a-d0cf-4465-a8a3-c52f3f5429c8 req-89c005d4-88f7-4712-999c-ccb7f98f6aad service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Refreshing network info cache for port b12a55b6-c396-44ab-a9ef-b183e97189e3 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 832.087686] env[62914]: DEBUG oslo_vmware.api [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352478, 'name': PowerOnVM_Task, 'duration_secs': 0.991485} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.087996] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 832.088158] env[62914]: INFO nova.compute.manager [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Took 9.56 seconds to spawn the instance on the hypervisor. [ 832.088338] env[62914]: DEBUG nova.compute.manager [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.089145] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce9caac0-3d76-4a91-ad0c-6e218b9774b0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.276955] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352479, 'name': ReconfigVM_Task, 'duration_secs': 0.665752} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.276955] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 46d139a1-a4ae-435e-9d6f-cfc06d706128/46d139a1-a4ae-435e-9d6f-cfc06d706128.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 832.276955] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d8c820f7-8c6a-4deb-82ab-f3afcab12eaf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.281600] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Waiting for the task: (returnval){ [ 832.281600] env[62914]: value = "task-1352481" [ 832.281600] env[62914]: _type = "Task" [ 832.281600] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.292430] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352481, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.333041] env[62914]: DEBUG nova.compute.utils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 832.333041] env[62914]: DEBUG nova.compute.manager [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 832.333041] env[62914]: DEBUG nova.network.neutron [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 832.386141] env[62914]: DEBUG nova.policy [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c9cb77e1af6b4f04b5c45072cc028210', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e47584f07144ba3ba9d662d98f2d043', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 832.393360] env[62914]: DEBUG nova.compute.manager [req-06a9a3bb-6fb3-4dcc-832a-8df6ddac4408 req-0000392d-e1cf-4a20-ade9-d9b59f7fdb99 service nova] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Received event network-vif-plugged-68eeb848-a0dd-455b-8501-a7b14519e427 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.394328] env[62914]: DEBUG oslo_concurrency.lockutils [req-06a9a3bb-6fb3-4dcc-832a-8df6ddac4408 req-0000392d-e1cf-4a20-ade9-d9b59f7fdb99 service nova] Acquiring lock "9577394d-1c73-4ed1-ba86-e7c246e32719-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.394711] env[62914]: DEBUG oslo_concurrency.lockutils [req-06a9a3bb-6fb3-4dcc-832a-8df6ddac4408 req-0000392d-e1cf-4a20-ade9-d9b59f7fdb99 service nova] Lock "9577394d-1c73-4ed1-ba86-e7c246e32719-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.395208] env[62914]: DEBUG oslo_concurrency.lockutils [req-06a9a3bb-6fb3-4dcc-832a-8df6ddac4408 req-0000392d-e1cf-4a20-ade9-d9b59f7fdb99 service nova] Lock "9577394d-1c73-4ed1-ba86-e7c246e32719-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.395706] env[62914]: DEBUG nova.compute.manager [req-06a9a3bb-6fb3-4dcc-832a-8df6ddac4408 req-0000392d-e1cf-4a20-ade9-d9b59f7fdb99 service nova] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] No waiting events found dispatching network-vif-plugged-68eeb848-a0dd-455b-8501-a7b14519e427 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 832.396129] env[62914]: WARNING nova.compute.manager [req-06a9a3bb-6fb3-4dcc-832a-8df6ddac4408 req-0000392d-e1cf-4a20-ade9-d9b59f7fdb99 service nova] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Received unexpected event network-vif-plugged-68eeb848-a0dd-455b-8501-a7b14519e427 for instance with vm_state building and task_state spawning. [ 832.474375] env[62914]: DEBUG oslo_vmware.api [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352480, 'name': PowerOffVM_Task, 'duration_secs': 0.17725} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.474826] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.475174] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 832.475306] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b41efc7-db47-475d-a272-4f193b08c9dd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.499671] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 832.499886] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 832.500130] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Deleting the datastore file [datastore1] 87ef6b06-f699-4de3-8b89-854717074406 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.500463] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cba5e42f-db0a-4286-b632-ee2caabfa1de {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.511696] env[62914]: DEBUG oslo_vmware.api [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for the task: (returnval){ [ 832.511696] env[62914]: value = "task-1352483" [ 832.511696] env[62914]: _type = "Task" [ 832.511696] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.519552] env[62914]: DEBUG oslo_vmware.api [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352483, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.530804] env[62914]: DEBUG nova.network.neutron [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Successfully updated port: 68eeb848-a0dd-455b-8501-a7b14519e427 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 832.536771] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Releasing lock "refresh_cache-cb89c84f-414b-4dc7-9db5-12cbc30b52b4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.536771] env[62914]: DEBUG nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Instance network_info: |[{"id": "1c71fa9c-a921-4649-81d3-9de5b41619c6", "address": "fa:16:3e:db:9a:a4", "network": {"id": "b3def210-2cac-4154-83ca-95b77b5e945c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1178445589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b6b4c54c48514783a044a75dcdea02b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c71fa9c-a9", "ovs_interfaceid": "1c71fa9c-a921-4649-81d3-9de5b41619c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 832.537089] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:9a:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50171613-b419-45e3-9ada-fcb6cd921428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1c71fa9c-a921-4649-81d3-9de5b41619c6', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 832.545972] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Creating folder: Project (b6b4c54c48514783a044a75dcdea02b8). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 832.546555] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-810e1187-f1e2-4516-bc28-778c6b3c69bb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.556483] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Created folder: Project (b6b4c54c48514783a044a75dcdea02b8) in parent group-v288131. [ 832.556722] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Creating folder: Instances. Parent ref: group-v288176. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 832.557685] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-109d9d7f-dbd6-43cb-a033-7cacc224360b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.571134] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Created folder: Instances in parent group-v288176. [ 832.571497] env[62914]: DEBUG oslo.service.loopingcall [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.577171] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 832.577171] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b301c404-f7a1-4868-8b09-171dfa201008 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.610505] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 832.610505] env[62914]: value = "task-1352486" [ 832.610505] env[62914]: _type = "Task" [ 832.610505] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.611070] env[62914]: INFO nova.compute.manager [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Took 28.42 seconds to build instance. [ 832.623590] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352486, 'name': CreateVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.791913] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352481, 'name': Rename_Task, 'duration_secs': 0.297222} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.796664] env[62914]: DEBUG nova.network.neutron [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Successfully created port: 9fadd240-af76-45c7-9f2b-f48b581f1362 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.798596] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 832.798861] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73a2b8fb-882f-4dcf-b1e4-4fa2b561c196 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.805740] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Waiting for the task: (returnval){ [ 832.805740] env[62914]: value = "task-1352487" [ 832.805740] env[62914]: _type = "Task" [ 832.805740] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.813780] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352487, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.834794] env[62914]: DEBUG nova.compute.manager [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 832.932403] env[62914]: DEBUG nova.network.neutron [req-bb8f0a1a-d0cf-4465-a8a3-c52f3f5429c8 req-89c005d4-88f7-4712-999c-ccb7f98f6aad service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Updated VIF entry in instance network info cache for port b12a55b6-c396-44ab-a9ef-b183e97189e3. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 832.932917] env[62914]: DEBUG nova.network.neutron [req-bb8f0a1a-d0cf-4465-a8a3-c52f3f5429c8 req-89c005d4-88f7-4712-999c-ccb7f98f6aad service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Updating instance_info_cache with network_info: [{"id": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "address": "fa:16:3e:7a:f0:fa", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb12a55b6-c3", "ovs_interfaceid": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.025085] env[62914]: DEBUG oslo_vmware.api [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Task: {'id': task-1352483, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139253} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.025469] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 833.025732] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 833.025942] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 833.026178] env[62914]: INFO nova.compute.manager [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Took 1.09 seconds to destroy the instance on the hypervisor. [ 833.026526] env[62914]: DEBUG oslo.service.loopingcall [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.026802] env[62914]: DEBUG nova.compute.manager [-] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 833.026947] env[62914]: DEBUG nova.network.neutron [-] [instance: 87ef6b06-f699-4de3-8b89-854717074406] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 833.040044] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "refresh_cache-9577394d-1c73-4ed1-ba86-e7c246e32719" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.040044] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquired lock "refresh_cache-9577394d-1c73-4ed1-ba86-e7c246e32719" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.040044] env[62914]: DEBUG nova.network.neutron [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.052379] env[62914]: DEBUG nova.network.neutron [-] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.105028] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e10a506a-1cc7-4120-8bfd-7d8d17869a65 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.116041] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be50319-c415-4c2b-b53a-4d2df65f3b94 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.120576] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f5b9e2bd-3526-43b1-8980-8ac659465d31 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "9fec3e97-4203-4a90-8a43-600d11d7e7a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.074s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.125839] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352486, 'name': CreateVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.151512] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-710723a8-41e1-4c41-97a3-bf4bf98915e8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.159189] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab041c81-680c-4ed5-b0b0-9afe21261fd1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.173876] env[62914]: DEBUG nova.compute.provider_tree [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.237159] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "9fec3e97-4203-4a90-8a43-600d11d7e7a0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.237447] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "9fec3e97-4203-4a90-8a43-600d11d7e7a0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.237654] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "9fec3e97-4203-4a90-8a43-600d11d7e7a0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.237834] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "9fec3e97-4203-4a90-8a43-600d11d7e7a0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.237997] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "9fec3e97-4203-4a90-8a43-600d11d7e7a0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.240235] env[62914]: INFO nova.compute.manager [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Terminating instance [ 833.242091] env[62914]: DEBUG nova.compute.manager [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 833.242286] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 833.243106] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd285ef-c1fa-4da5-b408-42100a8dc366 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.250976] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 833.251211] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-710537d5-1663-42a7-b386-4b663d9f293a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.257660] env[62914]: DEBUG oslo_vmware.api [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 833.257660] env[62914]: value = "task-1352488" [ 833.257660] env[62914]: _type = "Task" [ 833.257660] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.265626] env[62914]: DEBUG oslo_vmware.api [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352488, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.317874] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352487, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.441683] env[62914]: DEBUG oslo_concurrency.lockutils [req-bb8f0a1a-d0cf-4465-a8a3-c52f3f5429c8 req-89c005d4-88f7-4712-999c-ccb7f98f6aad service nova] Releasing lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.555032] env[62914]: DEBUG nova.network.neutron [-] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.571182] env[62914]: DEBUG nova.network.neutron [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.623693] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352486, 'name': CreateVM_Task, 'duration_secs': 0.543757} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.625929] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 833.626688] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.626858] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.627186] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 833.627878] env[62914]: DEBUG nova.compute.manager [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 833.630568] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6ac768b-eeb0-42c4-8b79-1fe91a5393a0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.636071] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 833.636071] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52faa285-2ff5-5f81-2c63-7b35dc4a98ca" [ 833.636071] env[62914]: _type = "Task" [ 833.636071] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.644562] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52faa285-2ff5-5f81-2c63-7b35dc4a98ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.676506] env[62914]: DEBUG nova.scheduler.client.report [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.758839] env[62914]: DEBUG nova.network.neutron [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Updating instance_info_cache with network_info: [{"id": "68eeb848-a0dd-455b-8501-a7b14519e427", "address": "fa:16:3e:76:5f:83", "network": {"id": "b3def210-2cac-4154-83ca-95b77b5e945c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1178445589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b6b4c54c48514783a044a75dcdea02b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68eeb848-a0", "ovs_interfaceid": "68eeb848-a0dd-455b-8501-a7b14519e427", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.769557] env[62914]: DEBUG oslo_vmware.api [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352488, 'name': PowerOffVM_Task, 'duration_secs': 0.187254} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.769812] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 833.770060] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 833.770285] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a75307b8-2399-4222-901f-33d77e6c8f68 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.816680] env[62914]: DEBUG oslo_vmware.api [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352487, 'name': PowerOnVM_Task, 'duration_secs': 0.733642} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.816991] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.817310] env[62914]: INFO nova.compute.manager [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Took 8.92 seconds to spawn the instance on the hypervisor. [ 833.817612] env[62914]: DEBUG nova.compute.manager [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.818318] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-147cb9aa-c5cc-4281-b901-1d1c64c7dc1d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.845568] env[62914]: DEBUG nova.compute.manager [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 833.848055] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 833.848055] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 833.848195] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleting the datastore file [datastore1] 9fec3e97-4203-4a90-8a43-600d11d7e7a0 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 833.848714] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3b93f73-70be-4463-b77f-0170c1a38a9f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.855038] env[62914]: DEBUG oslo_vmware.api [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 833.855038] env[62914]: value = "task-1352490" [ 833.855038] env[62914]: _type = "Task" [ 833.855038] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.863872] env[62914]: DEBUG oslo_vmware.api [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352490, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.873278] env[62914]: DEBUG nova.virt.hardware [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.873536] env[62914]: DEBUG nova.virt.hardware [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.873706] env[62914]: DEBUG nova.virt.hardware [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.873910] env[62914]: DEBUG nova.virt.hardware [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.874103] env[62914]: DEBUG nova.virt.hardware [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.874286] env[62914]: DEBUG nova.virt.hardware [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.874510] env[62914]: DEBUG nova.virt.hardware [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.874673] env[62914]: DEBUG nova.virt.hardware [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.874855] env[62914]: DEBUG nova.virt.hardware [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.875037] env[62914]: DEBUG nova.virt.hardware [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.875229] env[62914]: DEBUG nova.virt.hardware [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.876319] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6755a74b-c340-4364-abb0-248b04a5de23 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.883647] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8abb4cc-9f8d-40a2-9d42-191ff24238e4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.058035] env[62914]: INFO nova.compute.manager [-] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Took 1.03 seconds to deallocate network for instance. [ 834.146439] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52faa285-2ff5-5f81-2c63-7b35dc4a98ca, 'name': SearchDatastore_Task, 'duration_secs': 0.019747} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.146739] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.146957] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.147239] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.147390] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.147568] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.147816] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da405db6-b356-4aa1-b808-dbcfab8bde51 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.150178] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.156639] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.156813] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 834.157531] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2657727b-e819-467f-8ab2-e864878271dd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.162736] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 834.162736] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]529ccf33-2129-759c-71d6-c1560513b2aa" [ 834.162736] env[62914]: _type = "Task" [ 834.162736] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.170924] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]529ccf33-2129-759c-71d6-c1560513b2aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.182777] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.183278] env[62914]: DEBUG nova.compute.manager [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 834.185745] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.839s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.187213] env[62914]: INFO nova.compute.claims [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.264825] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Releasing lock "refresh_cache-9577394d-1c73-4ed1-ba86-e7c246e32719" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.265211] env[62914]: DEBUG nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Instance network_info: |[{"id": "68eeb848-a0dd-455b-8501-a7b14519e427", "address": "fa:16:3e:76:5f:83", "network": {"id": "b3def210-2cac-4154-83ca-95b77b5e945c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1178445589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b6b4c54c48514783a044a75dcdea02b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68eeb848-a0", "ovs_interfaceid": "68eeb848-a0dd-455b-8501-a7b14519e427", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 834.265596] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:5f:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50171613-b419-45e3-9ada-fcb6cd921428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '68eeb848-a0dd-455b-8501-a7b14519e427', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 834.273169] env[62914]: DEBUG oslo.service.loopingcall [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.273403] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 834.273600] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-adca8159-b983-4941-b1c1-1d706c698f22 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.295025] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 834.295025] env[62914]: value = "task-1352491" [ 834.295025] env[62914]: _type = "Task" [ 834.295025] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.302262] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352491, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.304341] env[62914]: DEBUG nova.compute.manager [req-2dd07614-a735-4d69-a374-61d26ca4505e req-6feccb7c-27f2-4f5a-a73a-0968b5e2f3aa service nova] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Received event network-vif-plugged-9fadd240-af76-45c7-9f2b-f48b581f1362 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 834.304556] env[62914]: DEBUG oslo_concurrency.lockutils [req-2dd07614-a735-4d69-a374-61d26ca4505e req-6feccb7c-27f2-4f5a-a73a-0968b5e2f3aa service nova] Acquiring lock "93d420a1-6d8f-4919-a42f-55aebab853ae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.304741] env[62914]: DEBUG oslo_concurrency.lockutils [req-2dd07614-a735-4d69-a374-61d26ca4505e req-6feccb7c-27f2-4f5a-a73a-0968b5e2f3aa service nova] Lock "93d420a1-6d8f-4919-a42f-55aebab853ae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.304904] env[62914]: DEBUG oslo_concurrency.lockutils [req-2dd07614-a735-4d69-a374-61d26ca4505e req-6feccb7c-27f2-4f5a-a73a-0968b5e2f3aa service nova] Lock "93d420a1-6d8f-4919-a42f-55aebab853ae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.305077] env[62914]: DEBUG nova.compute.manager [req-2dd07614-a735-4d69-a374-61d26ca4505e req-6feccb7c-27f2-4f5a-a73a-0968b5e2f3aa service nova] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] No waiting events found dispatching network-vif-plugged-9fadd240-af76-45c7-9f2b-f48b581f1362 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 834.305248] env[62914]: WARNING nova.compute.manager [req-2dd07614-a735-4d69-a374-61d26ca4505e req-6feccb7c-27f2-4f5a-a73a-0968b5e2f3aa service nova] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Received unexpected event network-vif-plugged-9fadd240-af76-45c7-9f2b-f48b581f1362 for instance with vm_state building and task_state spawning. [ 834.335114] env[62914]: INFO nova.compute.manager [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Took 25.97 seconds to build instance. [ 834.365440] env[62914]: DEBUG oslo_vmware.api [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352490, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200318} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.365701] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.365880] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 834.366084] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 834.366240] env[62914]: INFO nova.compute.manager [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Took 1.12 seconds to destroy the instance on the hypervisor. [ 834.366493] env[62914]: DEBUG oslo.service.loopingcall [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.366662] env[62914]: DEBUG nova.compute.manager [-] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.366757] env[62914]: DEBUG nova.network.neutron [-] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 834.390247] env[62914]: DEBUG nova.network.neutron [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Successfully updated port: 9fadd240-af76-45c7-9f2b-f48b581f1362 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 834.426116] env[62914]: DEBUG nova.compute.manager [req-6843b03f-25ef-4903-9005-841a5d518fe5 req-0831a17b-c270-4061-992c-2b1d1642f0d8 service nova] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Received event network-changed-68eeb848-a0dd-455b-8501-a7b14519e427 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 834.426330] env[62914]: DEBUG nova.compute.manager [req-6843b03f-25ef-4903-9005-841a5d518fe5 req-0831a17b-c270-4061-992c-2b1d1642f0d8 service nova] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Refreshing instance network info cache due to event network-changed-68eeb848-a0dd-455b-8501-a7b14519e427. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 834.426541] env[62914]: DEBUG oslo_concurrency.lockutils [req-6843b03f-25ef-4903-9005-841a5d518fe5 req-0831a17b-c270-4061-992c-2b1d1642f0d8 service nova] Acquiring lock "refresh_cache-9577394d-1c73-4ed1-ba86-e7c246e32719" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.426683] env[62914]: DEBUG oslo_concurrency.lockutils [req-6843b03f-25ef-4903-9005-841a5d518fe5 req-0831a17b-c270-4061-992c-2b1d1642f0d8 service nova] Acquired lock "refresh_cache-9577394d-1c73-4ed1-ba86-e7c246e32719" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.426838] env[62914]: DEBUG nova.network.neutron [req-6843b03f-25ef-4903-9005-841a5d518fe5 req-0831a17b-c270-4061-992c-2b1d1642f0d8 service nova] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Refreshing network info cache for port 68eeb848-a0dd-455b-8501-a7b14519e427 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 834.564959] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.672588] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]529ccf33-2129-759c-71d6-c1560513b2aa, 'name': SearchDatastore_Task, 'duration_secs': 0.010491} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.673348] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6de7c6c8-4707-4876-a60e-fcf51d03a1f1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.678028] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 834.678028] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]522652ac-2249-4226-8f54-7bd3d76100c8" [ 834.678028] env[62914]: _type = "Task" [ 834.678028] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.687357] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]522652ac-2249-4226-8f54-7bd3d76100c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.692148] env[62914]: DEBUG nova.compute.utils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 834.696310] env[62914]: DEBUG nova.compute.manager [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 834.696625] env[62914]: DEBUG nova.network.neutron [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 834.759627] env[62914]: DEBUG nova.policy [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a403e7bf5b0840cc9a64a2f0bb459124', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a585f45cffa4479a7d2391da6dc2e7c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 834.804982] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352491, 'name': CreateVM_Task, 'duration_secs': 0.337008} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.805185] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 834.805951] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.806126] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.806344] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.806629] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc561158-311f-48f0-b910-dc9d91412e7c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.815019] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 834.815019] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52dcd7af-4e4d-6cfb-fdf8-3aa19e4cb4ab" [ 834.815019] env[62914]: _type = "Task" [ 834.815019] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.819070] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52dcd7af-4e4d-6cfb-fdf8-3aa19e4cb4ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.836588] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3a5be120-7df4-4ac0-9357-cf7d98cf64dc tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "46d139a1-a4ae-435e-9d6f-cfc06d706128" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.477s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.892622] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Acquiring lock "refresh_cache-93d420a1-6d8f-4919-a42f-55aebab853ae" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.892773] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Acquired lock "refresh_cache-93d420a1-6d8f-4919-a42f-55aebab853ae" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.892918] env[62914]: DEBUG nova.network.neutron [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 835.054788] env[62914]: DEBUG nova.network.neutron [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Successfully created port: 0f287a59-bbed-4541-988e-f8481710291e {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 835.111918] env[62914]: DEBUG nova.network.neutron [-] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.193140] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]522652ac-2249-4226-8f54-7bd3d76100c8, 'name': SearchDatastore_Task, 'duration_secs': 0.012108} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.193140] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.193140] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] cb89c84f-414b-4dc7-9db5-12cbc30b52b4/cb89c84f-414b-4dc7-9db5-12cbc30b52b4.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 835.193140] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d1799f2-0e0f-4c41-b25a-ca1dbd963910 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.194381] env[62914]: DEBUG nova.compute.manager [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 835.197722] env[62914]: DEBUG nova.network.neutron [req-6843b03f-25ef-4903-9005-841a5d518fe5 req-0831a17b-c270-4061-992c-2b1d1642f0d8 service nova] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Updated VIF entry in instance network info cache for port 68eeb848-a0dd-455b-8501-a7b14519e427. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 835.198160] env[62914]: DEBUG nova.network.neutron [req-6843b03f-25ef-4903-9005-841a5d518fe5 req-0831a17b-c270-4061-992c-2b1d1642f0d8 service nova] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Updating instance_info_cache with network_info: [{"id": "68eeb848-a0dd-455b-8501-a7b14519e427", "address": "fa:16:3e:76:5f:83", "network": {"id": "b3def210-2cac-4154-83ca-95b77b5e945c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1178445589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b6b4c54c48514783a044a75dcdea02b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68eeb848-a0", "ovs_interfaceid": "68eeb848-a0dd-455b-8501-a7b14519e427", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.209579] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 835.209579] env[62914]: value = "task-1352492" [ 835.209579] env[62914]: _type = "Task" [ 835.209579] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.221000] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352492, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.329463] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52dcd7af-4e4d-6cfb-fdf8-3aa19e4cb4ab, 'name': SearchDatastore_Task, 'duration_secs': 0.009542} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.330109] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.330518] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.330879] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.331188] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.331686] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.332100] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c2bceee-4161-4848-8dcd-d5f977ed5c93 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.339799] env[62914]: DEBUG nova.compute.manager [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 835.343634] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.343939] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 835.346786] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4041b66-957c-45b6-ad69-7d1e01b3e5e3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.353645] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 835.353645] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]522541aa-90ee-1054-5418-a0a94d82b43e" [ 835.353645] env[62914]: _type = "Task" [ 835.353645] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.362785] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]522541aa-90ee-1054-5418-a0a94d82b43e, 'name': SearchDatastore_Task, 'duration_secs': 0.009369} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.366202] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-537dc3c8-5f4d-4f05-85ab-29d3bc75f5c0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.370901] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 835.370901] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ecf91e-64d8-9d71-a3e9-bc5bc5d00a83" [ 835.370901] env[62914]: _type = "Task" [ 835.370901] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.379222] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ecf91e-64d8-9d71-a3e9-bc5bc5d00a83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.432477] env[62914]: DEBUG nova.network.neutron [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.467993] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba1347e0-f151-4296-ac61-9826f546a00d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.475528] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662f7773-dc3a-4552-9b31-36967ee26443 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.514379] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef17733-8050-47bf-a505-b1ac9d25ebe4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.523075] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a1bc5d-d4de-4e40-9c2b-c7a1b88859ad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.538550] env[62914]: DEBUG nova.compute.provider_tree [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.614803] env[62914]: INFO nova.compute.manager [-] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Took 1.25 seconds to deallocate network for instance. [ 835.655574] env[62914]: DEBUG nova.network.neutron [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Updating instance_info_cache with network_info: [{"id": "9fadd240-af76-45c7-9f2b-f48b581f1362", "address": "fa:16:3e:03:a5:89", "network": {"id": "65bb9439-a585-4b59-8269-d0f2f2701fa0", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-554325180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e47584f07144ba3ba9d662d98f2d043", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fadd240-af", "ovs_interfaceid": "9fadd240-af76-45c7-9f2b-f48b581f1362", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.709819] env[62914]: DEBUG oslo_concurrency.lockutils [req-6843b03f-25ef-4903-9005-841a5d518fe5 req-0831a17b-c270-4061-992c-2b1d1642f0d8 service nova] Releasing lock "refresh_cache-9577394d-1c73-4ed1-ba86-e7c246e32719" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.727245] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352492, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.861860] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.880776] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ecf91e-64d8-9d71-a3e9-bc5bc5d00a83, 'name': SearchDatastore_Task, 'duration_secs': 0.009472} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.881202] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.881565] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 9577394d-1c73-4ed1-ba86-e7c246e32719/9577394d-1c73-4ed1-ba86-e7c246e32719.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 835.881958] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b3a3d407-064e-4aa1-9e85-93906855422d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.891521] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 835.891521] env[62914]: value = "task-1352493" [ 835.891521] env[62914]: _type = "Task" [ 835.891521] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.899396] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352493, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.042039] env[62914]: DEBUG nova.scheduler.client.report [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.122160] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.158394] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Releasing lock "refresh_cache-93d420a1-6d8f-4919-a42f-55aebab853ae" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.158743] env[62914]: DEBUG nova.compute.manager [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Instance network_info: |[{"id": "9fadd240-af76-45c7-9f2b-f48b581f1362", "address": "fa:16:3e:03:a5:89", "network": {"id": "65bb9439-a585-4b59-8269-d0f2f2701fa0", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-554325180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e47584f07144ba3ba9d662d98f2d043", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fadd240-af", "ovs_interfaceid": "9fadd240-af76-45c7-9f2b-f48b581f1362", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 836.159259] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:a5:89', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7b2049d7-f99e-425a-afdb-2c95ca88e483', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9fadd240-af76-45c7-9f2b-f48b581f1362', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 836.166964] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Creating folder: Project (9e47584f07144ba3ba9d662d98f2d043). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 836.167273] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-02db18ea-3640-48a3-a9e8-b3a6d8d30e9a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.177862] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "46d139a1-a4ae-435e-9d6f-cfc06d706128" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.178249] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "46d139a1-a4ae-435e-9d6f-cfc06d706128" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.178394] env[62914]: INFO nova.compute.manager [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Rebooting instance [ 836.179725] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Created folder: Project (9e47584f07144ba3ba9d662d98f2d043) in parent group-v288131. [ 836.179899] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Creating folder: Instances. Parent ref: group-v288180. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 836.180358] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e573f6f0-118a-4b51-8e17-9a27a48aa187 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.189907] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Created folder: Instances in parent group-v288180. [ 836.190165] env[62914]: DEBUG oslo.service.loopingcall [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.190363] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 836.191157] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fbfaed9a-bb23-40c2-9160-09d4c5f10144 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.210901] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 836.210901] env[62914]: value = "task-1352496" [ 836.210901] env[62914]: _type = "Task" [ 836.210901] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.220847] env[62914]: DEBUG nova.compute.manager [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 836.222600] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352496, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.227640] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352492, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.594205} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.227871] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] cb89c84f-414b-4dc7-9db5-12cbc30b52b4/cb89c84f-414b-4dc7-9db5-12cbc30b52b4.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 836.228097] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.228325] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e0ec694c-1af4-4f50-ae2d-f8cfa0544055 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.234387] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 836.234387] env[62914]: value = "task-1352497" [ 836.234387] env[62914]: _type = "Task" [ 836.234387] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.243569] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352497, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.248941] env[62914]: DEBUG nova.virt.hardware [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.249233] env[62914]: DEBUG nova.virt.hardware [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.249399] env[62914]: DEBUG nova.virt.hardware [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.249608] env[62914]: DEBUG nova.virt.hardware [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.249759] env[62914]: DEBUG nova.virt.hardware [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.249905] env[62914]: DEBUG nova.virt.hardware [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.250164] env[62914]: DEBUG nova.virt.hardware [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.250360] env[62914]: DEBUG nova.virt.hardware [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.250518] env[62914]: DEBUG nova.virt.hardware [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.250707] env[62914]: DEBUG nova.virt.hardware [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.250868] env[62914]: DEBUG nova.virt.hardware [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.251792] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c397db37-e5d5-4a6e-b12c-1b606afca6ac {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.260188] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e16e259-0b84-49e6-bb9d-732f1af8ead0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.337283] env[62914]: DEBUG nova.compute.manager [req-bd918170-ec43-4b26-a796-6d26b529c03a req-4e68802c-6c8d-40f1-80f6-3eecc3db7aa6 service nova] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Received event network-changed-9fadd240-af76-45c7-9f2b-f48b581f1362 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.337476] env[62914]: DEBUG nova.compute.manager [req-bd918170-ec43-4b26-a796-6d26b529c03a req-4e68802c-6c8d-40f1-80f6-3eecc3db7aa6 service nova] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Refreshing instance network info cache due to event network-changed-9fadd240-af76-45c7-9f2b-f48b581f1362. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 836.337692] env[62914]: DEBUG oslo_concurrency.lockutils [req-bd918170-ec43-4b26-a796-6d26b529c03a req-4e68802c-6c8d-40f1-80f6-3eecc3db7aa6 service nova] Acquiring lock "refresh_cache-93d420a1-6d8f-4919-a42f-55aebab853ae" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.337914] env[62914]: DEBUG oslo_concurrency.lockutils [req-bd918170-ec43-4b26-a796-6d26b529c03a req-4e68802c-6c8d-40f1-80f6-3eecc3db7aa6 service nova] Acquired lock "refresh_cache-93d420a1-6d8f-4919-a42f-55aebab853ae" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.337983] env[62914]: DEBUG nova.network.neutron [req-bd918170-ec43-4b26-a796-6d26b529c03a req-4e68802c-6c8d-40f1-80f6-3eecc3db7aa6 service nova] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Refreshing network info cache for port 9fadd240-af76-45c7-9f2b-f48b581f1362 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 836.405747] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352493, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.478826] env[62914]: DEBUG nova.compute.manager [req-edbb7226-6ecf-4ec1-b1e1-de2c65bd3fc0 req-e08d7044-82c2-4460-9db2-b4cc71586032 service nova] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Received event network-vif-deleted-b82c2ddd-9dd3-4bd8-ad01-68787d570158 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.479281] env[62914]: DEBUG nova.compute.manager [req-edbb7226-6ecf-4ec1-b1e1-de2c65bd3fc0 req-e08d7044-82c2-4460-9db2-b4cc71586032 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Received event network-changed-4f7e867d-96a4-4d4e-82fa-43469303566c {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.479667] env[62914]: DEBUG nova.compute.manager [req-edbb7226-6ecf-4ec1-b1e1-de2c65bd3fc0 req-e08d7044-82c2-4460-9db2-b4cc71586032 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Refreshing instance network info cache due to event network-changed-4f7e867d-96a4-4d4e-82fa-43469303566c. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 836.479949] env[62914]: DEBUG oslo_concurrency.lockutils [req-edbb7226-6ecf-4ec1-b1e1-de2c65bd3fc0 req-e08d7044-82c2-4460-9db2-b4cc71586032 service nova] Acquiring lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.480289] env[62914]: DEBUG oslo_concurrency.lockutils [req-edbb7226-6ecf-4ec1-b1e1-de2c65bd3fc0 req-e08d7044-82c2-4460-9db2-b4cc71586032 service nova] Acquired lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.480546] env[62914]: DEBUG nova.network.neutron [req-edbb7226-6ecf-4ec1-b1e1-de2c65bd3fc0 req-e08d7044-82c2-4460-9db2-b4cc71586032 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Refreshing network info cache for port 4f7e867d-96a4-4d4e-82fa-43469303566c {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 836.546709] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.361s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.547319] env[62914]: DEBUG nova.compute.manager [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 836.549986] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.487s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.551521] env[62914]: INFO nova.compute.claims [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.701701] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.721725] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352496, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.743871] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352497, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091092} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.744229] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.745025] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2cd670a-9a28-49ae-8f15-38d4eec2ff7b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.768914] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] cb89c84f-414b-4dc7-9db5-12cbc30b52b4/cb89c84f-414b-4dc7-9db5-12cbc30b52b4.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.769882] env[62914]: DEBUG nova.network.neutron [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Successfully updated port: 0f287a59-bbed-4541-988e-f8481710291e {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 836.771186] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea587b6b-5b14-4e1c-8131-f106c665368e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.792457] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 836.792457] env[62914]: value = "task-1352498" [ 836.792457] env[62914]: _type = "Task" [ 836.792457] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.800667] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352498, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.901503] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352493, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.695077} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.901769] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 9577394d-1c73-4ed1-ba86-e7c246e32719/9577394d-1c73-4ed1-ba86-e7c246e32719.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 836.902015] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.902232] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b61aaf0b-cf6d-479d-9070-f88d1740fb3f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.910490] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 836.910490] env[62914]: value = "task-1352499" [ 836.910490] env[62914]: _type = "Task" [ 836.910490] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.918556] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352499, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.055996] env[62914]: DEBUG nova.compute.utils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.059601] env[62914]: DEBUG nova.compute.manager [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 837.059812] env[62914]: DEBUG nova.network.neutron [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 837.105867] env[62914]: DEBUG nova.network.neutron [req-bd918170-ec43-4b26-a796-6d26b529c03a req-4e68802c-6c8d-40f1-80f6-3eecc3db7aa6 service nova] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Updated VIF entry in instance network info cache for port 9fadd240-af76-45c7-9f2b-f48b581f1362. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 837.105867] env[62914]: DEBUG nova.network.neutron [req-bd918170-ec43-4b26-a796-6d26b529c03a req-4e68802c-6c8d-40f1-80f6-3eecc3db7aa6 service nova] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Updating instance_info_cache with network_info: [{"id": "9fadd240-af76-45c7-9f2b-f48b581f1362", "address": "fa:16:3e:03:a5:89", "network": {"id": "65bb9439-a585-4b59-8269-d0f2f2701fa0", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-554325180-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e47584f07144ba3ba9d662d98f2d043", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fadd240-af", "ovs_interfaceid": "9fadd240-af76-45c7-9f2b-f48b581f1362", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.122188] env[62914]: DEBUG nova.policy [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '795b94761b144a1e99ed79838d9687ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9278818cebd842669b802a4e43e91774', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 837.222898] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352496, 'name': CreateVM_Task, 'duration_secs': 0.771469} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.223760] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.224635] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.227029] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.227029] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.227029] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd597756-0cfb-4142-a52e-08eef5062a3b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.233127] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 837.233127] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52234931-c712-18c0-224c-d3af07d62b55" [ 837.233127] env[62914]: _type = "Task" [ 837.233127] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.246413] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52234931-c712-18c0-224c-d3af07d62b55, 'name': SearchDatastore_Task, 'duration_secs': 0.008641} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.246880] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.247292] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.247764] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.250018] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.250018] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.250018] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c510624e-aef0-4973-a196-90b878459a84 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.255081] env[62914]: DEBUG nova.network.neutron [req-edbb7226-6ecf-4ec1-b1e1-de2c65bd3fc0 req-e08d7044-82c2-4460-9db2-b4cc71586032 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Updated VIF entry in instance network info cache for port 4f7e867d-96a4-4d4e-82fa-43469303566c. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 837.255563] env[62914]: DEBUG nova.network.neutron [req-edbb7226-6ecf-4ec1-b1e1-de2c65bd3fc0 req-e08d7044-82c2-4460-9db2-b4cc71586032 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Updating instance_info_cache with network_info: [{"id": "4f7e867d-96a4-4d4e-82fa-43469303566c", "address": "fa:16:3e:51:58:7b", "network": {"id": "c8cdb39e-9415-4f67-8199-fa1b043d04d8", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-373473409-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f897e1c819f44362857de4d4398ca613", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f7e867d-96", "ovs_interfaceid": "4f7e867d-96a4-4d4e-82fa-43469303566c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.258476] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.258941] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 837.260531] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df140b37-dcfc-4a6c-8d3e-3f56a6192bcb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.267264] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 837.267264] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52bfcae9-11e7-53d1-b84e-37636e630061" [ 837.267264] env[62914]: _type = "Task" [ 837.267264] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.278395] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52bfcae9-11e7-53d1-b84e-37636e630061, 'name': SearchDatastore_Task, 'duration_secs': 0.009056} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.279249] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46f09f5e-bac3-44e5-bab7-9133ff37870a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.285490] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 837.285490] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52febac4-d259-0509-4266-71ec3a68fd36" [ 837.285490] env[62914]: _type = "Task" [ 837.285490] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.287896] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Acquiring lock "refresh_cache-bd3e82ce-62e7-4761-bcd7-d1f835234b97" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.287896] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Acquired lock "refresh_cache-bd3e82ce-62e7-4761-bcd7-d1f835234b97" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.288049] env[62914]: DEBUG nova.network.neutron [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 837.293576] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52febac4-d259-0509-4266-71ec3a68fd36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.301730] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352498, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.405310] env[62914]: DEBUG nova.network.neutron [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Successfully created port: a5b49567-42ef-4e73-ab33-113aa689e83b {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.420120] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352499, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.148942} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.420433] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 837.421203] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924a9964-7efd-4303-86fe-9d0303e70fff {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.443056] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 9577394d-1c73-4ed1-ba86-e7c246e32719/9577394d-1c73-4ed1-ba86-e7c246e32719.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.443407] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ac6d076-2196-4b58-a912-6e60d01ee779 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.463069] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 837.463069] env[62914]: value = "task-1352500" [ 837.463069] env[62914]: _type = "Task" [ 837.463069] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.470770] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352500, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.560518] env[62914]: DEBUG nova.compute.manager [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 837.607453] env[62914]: DEBUG oslo_concurrency.lockutils [req-bd918170-ec43-4b26-a796-6d26b529c03a req-4e68802c-6c8d-40f1-80f6-3eecc3db7aa6 service nova] Releasing lock "refresh_cache-93d420a1-6d8f-4919-a42f-55aebab853ae" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.761755] env[62914]: DEBUG oslo_concurrency.lockutils [req-edbb7226-6ecf-4ec1-b1e1-de2c65bd3fc0 req-e08d7044-82c2-4460-9db2-b4cc71586032 service nova] Releasing lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.762174] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquired lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.762363] env[62914]: DEBUG nova.network.neutron [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 837.774833] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9494156c-cdeb-48d0-9b95-0f99780f7d28 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.782138] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be5a919e-6891-461e-8f22-957b42fab532 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.819305] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52febac4-d259-0509-4266-71ec3a68fd36, 'name': SearchDatastore_Task, 'duration_secs': 0.01051} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.822796] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54823a8-3cef-4fd4-a7e8-6e1cc36c6c89 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.824964] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.825238] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 93d420a1-6d8f-4919-a42f-55aebab853ae/93d420a1-6d8f-4919-a42f-55aebab853ae.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 837.825477] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cacf8975-133d-45b5-84d1-53161cda6f38 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.832604] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352498, 'name': ReconfigVM_Task, 'duration_secs': 0.553148} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.835975] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Reconfigured VM instance instance-00000042 to attach disk [datastore2] cb89c84f-414b-4dc7-9db5-12cbc30b52b4/cb89c84f-414b-4dc7-9db5-12cbc30b52b4.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.836941] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee529946-0d69-4168-a848-2184a1e0eda6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.839285] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88924dd3-4664-4e5c-8f92-51c48a62ae4d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.843973] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 837.843973] env[62914]: value = "task-1352501" [ 837.843973] env[62914]: _type = "Task" [ 837.843973] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.845256] env[62914]: DEBUG nova.network.neutron [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.857865] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 837.857865] env[62914]: value = "task-1352502" [ 837.857865] env[62914]: _type = "Task" [ 837.857865] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.858570] env[62914]: DEBUG nova.compute.provider_tree [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.865977] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352501, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.870524] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352502, 'name': Rename_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.973812] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352500, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.021871] env[62914]: DEBUG nova.network.neutron [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Updating instance_info_cache with network_info: [{"id": "0f287a59-bbed-4541-988e-f8481710291e", "address": "fa:16:3e:8a:c2:de", "network": {"id": "96bdedee-1b1f-4e67-9718-03b96b0090c9", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1196764842-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a585f45cffa4479a7d2391da6dc2e7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "537e0890-4fa2-4f2d-b74c-49933a4edf53", "external-id": "nsx-vlan-transportzone-82", "segmentation_id": 82, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f287a59-bb", "ovs_interfaceid": "0f287a59-bbed-4541-988e-f8481710291e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.356417] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352501, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508953} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.356729] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 93d420a1-6d8f-4919-a42f-55aebab853ae/93d420a1-6d8f-4919-a42f-55aebab853ae.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 838.356919] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.357218] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c7d61cf-362e-466d-b7cb-846a48a173ab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.360996] env[62914]: DEBUG nova.scheduler.client.report [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.370413] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 838.370413] env[62914]: value = "task-1352503" [ 838.370413] env[62914]: _type = "Task" [ 838.370413] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.376807] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352502, 'name': Rename_Task, 'duration_secs': 0.129794} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.377892] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 838.378145] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f945c9eb-332e-4ea2-8f74-57c99b657230 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.382325] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352503, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.387195] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 838.387195] env[62914]: value = "task-1352504" [ 838.387195] env[62914]: _type = "Task" [ 838.387195] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.394547] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352504, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.475164] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352500, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.494616] env[62914]: DEBUG nova.network.neutron [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Updating instance_info_cache with network_info: [{"id": "4f7e867d-96a4-4d4e-82fa-43469303566c", "address": "fa:16:3e:51:58:7b", "network": {"id": "c8cdb39e-9415-4f67-8199-fa1b043d04d8", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-373473409-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f897e1c819f44362857de4d4398ca613", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f7e867d-96", "ovs_interfaceid": "4f7e867d-96a4-4d4e-82fa-43469303566c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.523307] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Releasing lock "refresh_cache-bd3e82ce-62e7-4761-bcd7-d1f835234b97" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.523610] env[62914]: DEBUG nova.compute.manager [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Instance network_info: |[{"id": "0f287a59-bbed-4541-988e-f8481710291e", "address": "fa:16:3e:8a:c2:de", "network": {"id": "96bdedee-1b1f-4e67-9718-03b96b0090c9", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1196764842-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a585f45cffa4479a7d2391da6dc2e7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "537e0890-4fa2-4f2d-b74c-49933a4edf53", "external-id": "nsx-vlan-transportzone-82", "segmentation_id": 82, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f287a59-bb", "ovs_interfaceid": "0f287a59-bbed-4541-988e-f8481710291e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 838.524022] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:c2:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '537e0890-4fa2-4f2d-b74c-49933a4edf53', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0f287a59-bbed-4541-988e-f8481710291e', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 838.531630] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Creating folder: Project (1a585f45cffa4479a7d2391da6dc2e7c). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 838.531899] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-830d0c70-905f-4e65-b6a4-0207f84cc933 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.543188] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Created folder: Project (1a585f45cffa4479a7d2391da6dc2e7c) in parent group-v288131. [ 838.543380] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Creating folder: Instances. Parent ref: group-v288183. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 838.543617] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-56c3366b-d2ce-4afd-a192-e28324e564ba {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.554105] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Created folder: Instances in parent group-v288183. [ 838.554361] env[62914]: DEBUG oslo.service.loopingcall [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.554578] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 838.554802] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2757430e-e4c1-4681-b5c7-c9da6b5a22bd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.574114] env[62914]: DEBUG nova.compute.manager [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 838.577424] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 838.577424] env[62914]: value = "task-1352507" [ 838.577424] env[62914]: _type = "Task" [ 838.577424] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.586028] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352507, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.601261] env[62914]: DEBUG nova.virt.hardware [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 838.601626] env[62914]: DEBUG nova.virt.hardware [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 838.601626] env[62914]: DEBUG nova.virt.hardware [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.601788] env[62914]: DEBUG nova.virt.hardware [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 838.601839] env[62914]: DEBUG nova.virt.hardware [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.601953] env[62914]: DEBUG nova.virt.hardware [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 838.602174] env[62914]: DEBUG nova.virt.hardware [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 838.602332] env[62914]: DEBUG nova.virt.hardware [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 838.602497] env[62914]: DEBUG nova.virt.hardware [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 838.602655] env[62914]: DEBUG nova.virt.hardware [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 838.602821] env[62914]: DEBUG nova.virt.hardware [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 838.603678] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-338078d7-4c54-47c7-9268-d6051d761d12 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.608649] env[62914]: DEBUG nova.compute.manager [req-3b6e3cfa-e8c7-42a6-a4dc-dd0f8dde047e req-d97b9851-f55b-4955-99db-06a57158a96e service nova] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Received event network-vif-plugged-0f287a59-bbed-4541-988e-f8481710291e {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.608733] env[62914]: DEBUG oslo_concurrency.lockutils [req-3b6e3cfa-e8c7-42a6-a4dc-dd0f8dde047e req-d97b9851-f55b-4955-99db-06a57158a96e service nova] Acquiring lock "bd3e82ce-62e7-4761-bcd7-d1f835234b97-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.608934] env[62914]: DEBUG oslo_concurrency.lockutils [req-3b6e3cfa-e8c7-42a6-a4dc-dd0f8dde047e req-d97b9851-f55b-4955-99db-06a57158a96e service nova] Lock "bd3e82ce-62e7-4761-bcd7-d1f835234b97-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.609138] env[62914]: DEBUG oslo_concurrency.lockutils [req-3b6e3cfa-e8c7-42a6-a4dc-dd0f8dde047e req-d97b9851-f55b-4955-99db-06a57158a96e service nova] Lock "bd3e82ce-62e7-4761-bcd7-d1f835234b97-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.609302] env[62914]: DEBUG nova.compute.manager [req-3b6e3cfa-e8c7-42a6-a4dc-dd0f8dde047e req-d97b9851-f55b-4955-99db-06a57158a96e service nova] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] No waiting events found dispatching network-vif-plugged-0f287a59-bbed-4541-988e-f8481710291e {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 838.609472] env[62914]: WARNING nova.compute.manager [req-3b6e3cfa-e8c7-42a6-a4dc-dd0f8dde047e req-d97b9851-f55b-4955-99db-06a57158a96e service nova] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Received unexpected event network-vif-plugged-0f287a59-bbed-4541-988e-f8481710291e for instance with vm_state building and task_state spawning. [ 838.609625] env[62914]: DEBUG nova.compute.manager [req-3b6e3cfa-e8c7-42a6-a4dc-dd0f8dde047e req-d97b9851-f55b-4955-99db-06a57158a96e service nova] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Received event network-changed-0f287a59-bbed-4541-988e-f8481710291e {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.609770] env[62914]: DEBUG nova.compute.manager [req-3b6e3cfa-e8c7-42a6-a4dc-dd0f8dde047e req-d97b9851-f55b-4955-99db-06a57158a96e service nova] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Refreshing instance network info cache due to event network-changed-0f287a59-bbed-4541-988e-f8481710291e. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.609946] env[62914]: DEBUG oslo_concurrency.lockutils [req-3b6e3cfa-e8c7-42a6-a4dc-dd0f8dde047e req-d97b9851-f55b-4955-99db-06a57158a96e service nova] Acquiring lock "refresh_cache-bd3e82ce-62e7-4761-bcd7-d1f835234b97" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.610090] env[62914]: DEBUG oslo_concurrency.lockutils [req-3b6e3cfa-e8c7-42a6-a4dc-dd0f8dde047e req-d97b9851-f55b-4955-99db-06a57158a96e service nova] Acquired lock "refresh_cache-bd3e82ce-62e7-4761-bcd7-d1f835234b97" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.610288] env[62914]: DEBUG nova.network.neutron [req-3b6e3cfa-e8c7-42a6-a4dc-dd0f8dde047e req-d97b9851-f55b-4955-99db-06a57158a96e service nova] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Refreshing network info cache for port 0f287a59-bbed-4541-988e-f8481710291e {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 838.616689] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f31d3c3b-643a-4b45-b7c7-4125373a24b9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.870712] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.321s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.871261] env[62914]: DEBUG nova.compute.manager [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 838.874712] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.395s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.876321] env[62914]: INFO nova.compute.claims [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.889022] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352503, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064407} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.891872] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.892999] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4777d22a-ce44-4238-9d09-8d9fca17a49c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.916514] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 93d420a1-6d8f-4919-a42f-55aebab853ae/93d420a1-6d8f-4919-a42f-55aebab853ae.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.920370] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-114c7ce1-bac9-41bf-a7af-a03d651cfcab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.935722] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352504, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.941138] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 838.941138] env[62914]: value = "task-1352508" [ 838.941138] env[62914]: _type = "Task" [ 838.941138] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.949690] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352508, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.976750] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352500, 'name': ReconfigVM_Task, 'duration_secs': 1.210835} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.977706] env[62914]: DEBUG nova.network.neutron [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Successfully updated port: a5b49567-42ef-4e73-ab33-113aa689e83b {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 838.978887] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 9577394d-1c73-4ed1-ba86-e7c246e32719/9577394d-1c73-4ed1-ba86-e7c246e32719.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.979528] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3d55f4a-3e40-4fe4-afce-d79b2f1377fc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.986174] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 838.986174] env[62914]: value = "task-1352509" [ 838.986174] env[62914]: _type = "Task" [ 838.986174] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.997634] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Releasing lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.999339] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352509, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.999751] env[62914]: DEBUG nova.compute.manager [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.000567] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5beb2b-8bb6-4d79-a8a7-621b09761eaf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.088086] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352507, 'name': CreateVM_Task, 'duration_secs': 0.422608} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.088277] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 839.089050] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.089398] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.089772] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 839.091074] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3f804ed-ec43-45f2-9e9a-3058ac55db3e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.095061] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Waiting for the task: (returnval){ [ 839.095061] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]522fdc4e-d46d-9080-a297-7e710a579a76" [ 839.095061] env[62914]: _type = "Task" [ 839.095061] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.102881] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]522fdc4e-d46d-9080-a297-7e710a579a76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.368877] env[62914]: DEBUG nova.network.neutron [req-3b6e3cfa-e8c7-42a6-a4dc-dd0f8dde047e req-d97b9851-f55b-4955-99db-06a57158a96e service nova] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Updated VIF entry in instance network info cache for port 0f287a59-bbed-4541-988e-f8481710291e. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 839.369347] env[62914]: DEBUG nova.network.neutron [req-3b6e3cfa-e8c7-42a6-a4dc-dd0f8dde047e req-d97b9851-f55b-4955-99db-06a57158a96e service nova] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Updating instance_info_cache with network_info: [{"id": "0f287a59-bbed-4541-988e-f8481710291e", "address": "fa:16:3e:8a:c2:de", "network": {"id": "96bdedee-1b1f-4e67-9718-03b96b0090c9", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1196764842-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a585f45cffa4479a7d2391da6dc2e7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "537e0890-4fa2-4f2d-b74c-49933a4edf53", "external-id": "nsx-vlan-transportzone-82", "segmentation_id": 82, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f287a59-bb", "ovs_interfaceid": "0f287a59-bbed-4541-988e-f8481710291e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.381207] env[62914]: DEBUG nova.compute.utils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.387847] env[62914]: DEBUG nova.compute.manager [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 839.388080] env[62914]: DEBUG nova.network.neutron [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 839.399695] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352504, 'name': PowerOnVM_Task, 'duration_secs': 0.605188} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.399939] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 839.400152] env[62914]: INFO nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Took 10.33 seconds to spawn the instance on the hypervisor. [ 839.400338] env[62914]: DEBUG nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.401112] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e280286-3505-438c-aa75-a7fd84f7d55c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.431140] env[62914]: DEBUG nova.policy [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd900131bc974504a489fcd4ffaec5b6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c3055903b6f4443bbeb9897e75e34c49', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 839.451752] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352508, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.483640] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "refresh_cache-2274ca27-8e2d-435e-8570-97eb4ae3cc58" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.483640] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "refresh_cache-2274ca27-8e2d-435e-8570-97eb4ae3cc58" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.483640] env[62914]: DEBUG nova.network.neutron [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 839.496509] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352509, 'name': Rename_Task, 'duration_secs': 0.474205} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.496772] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 839.497151] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4eaba698-a228-49f9-aa6e-d04ff0fe269e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.503854] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 839.503854] env[62914]: value = "task-1352510" [ 839.503854] env[62914]: _type = "Task" [ 839.503854] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.513839] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352510, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.607133] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]522fdc4e-d46d-9080-a297-7e710a579a76, 'name': SearchDatastore_Task, 'duration_secs': 0.041856} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.607414] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.607648] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.607927] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.608090] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.608269] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.608530] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-45d709da-d65d-40df-8b91-d7a9d62dd72b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.619554] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.619722] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 839.620432] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08902699-be91-4728-8b04-ad3cecd1de87 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.625439] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Waiting for the task: (returnval){ [ 839.625439] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5256f460-1f3b-6dc3-f825-4b3d96def704" [ 839.625439] env[62914]: _type = "Task" [ 839.625439] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.633007] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5256f460-1f3b-6dc3-f825-4b3d96def704, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.683908] env[62914]: DEBUG nova.network.neutron [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Successfully created port: 59eaafaa-3160-4236-ab11-eb6f5e3ffec8 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 839.874301] env[62914]: DEBUG oslo_concurrency.lockutils [req-3b6e3cfa-e8c7-42a6-a4dc-dd0f8dde047e req-d97b9851-f55b-4955-99db-06a57158a96e service nova] Releasing lock "refresh_cache-bd3e82ce-62e7-4761-bcd7-d1f835234b97" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.895458] env[62914]: DEBUG nova.compute.manager [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 839.920647] env[62914]: INFO nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Took 29.65 seconds to build instance. [ 839.953958] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352508, 'name': ReconfigVM_Task, 'duration_secs': 0.964474} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.956912] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 93d420a1-6d8f-4919-a42f-55aebab853ae/93d420a1-6d8f-4919-a42f-55aebab853ae.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.957914] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ebdf61f2-7450-444f-95c8-fd3494b5da7c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.964625] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 839.964625] env[62914]: value = "task-1352511" [ 839.964625] env[62914]: _type = "Task" [ 839.964625] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.973890] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352511, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.013559] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352510, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.018475] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e2851e-a976-46c4-ac38-aeef2a86cc2b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.021570] env[62914]: DEBUG nova.network.neutron [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 840.027045] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Doing hard reboot of VM {{(pid=62914) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 840.027297] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-c4203e7f-de33-4fc8-9208-c54799cf595f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.033052] env[62914]: DEBUG oslo_vmware.api [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Waiting for the task: (returnval){ [ 840.033052] env[62914]: value = "task-1352512" [ 840.033052] env[62914]: _type = "Task" [ 840.033052] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.044904] env[62914]: DEBUG oslo_vmware.api [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352512, 'name': ResetVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.135259] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5256f460-1f3b-6dc3-f825-4b3d96def704, 'name': SearchDatastore_Task, 'duration_secs': 0.072419} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.137149] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4516efd-66a0-4e6c-9eff-32ebf1dba313 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.139523] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c67f826-e3af-4139-89f4-478aa7050cb7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.146887] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b442be-dddd-4bae-9baf-4951e9d61a94 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.150278] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Waiting for the task: (returnval){ [ 840.150278] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]520e9bfb-99dd-d660-bbe9-d3946d360821" [ 840.150278] env[62914]: _type = "Task" [ 840.150278] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.154501] env[62914]: DEBUG nova.network.neutron [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Updating instance_info_cache with network_info: [{"id": "a5b49567-42ef-4e73-ab33-113aa689e83b", "address": "fa:16:3e:d8:88:64", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5b49567-42", "ovs_interfaceid": "a5b49567-42ef-4e73-ab33-113aa689e83b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.182065] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "refresh_cache-2274ca27-8e2d-435e-8570-97eb4ae3cc58" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.182360] env[62914]: DEBUG nova.compute.manager [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Instance network_info: |[{"id": "a5b49567-42ef-4e73-ab33-113aa689e83b", "address": "fa:16:3e:d8:88:64", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5b49567-42", "ovs_interfaceid": "a5b49567-42ef-4e73-ab33-113aa689e83b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 840.183115] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4fe8638-e86c-4612-9ab1-6a36ed83ea1b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.185736] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:88:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '32faf59b-014c-4f1f-8331-40df95bf741f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a5b49567-42ef-4e73-ab33-113aa689e83b', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 840.192887] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Creating folder: Project (9278818cebd842669b802a4e43e91774). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 840.196964] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53b488f8-6329-4bbf-8a8f-e74907857336 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.198780] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]520e9bfb-99dd-d660-bbe9-d3946d360821, 'name': SearchDatastore_Task, 'duration_secs': 0.008695} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.199456] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.199688] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] bd3e82ce-62e7-4761-bcd7-d1f835234b97/bd3e82ce-62e7-4761-bcd7-d1f835234b97.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 840.200247] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-01b27d9c-12f8-45a3-ad4f-8e03f5434cd2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.207034] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1545790a-dee3-4e3b-ab4c-a3af696daeca {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.212364] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Created folder: Project (9278818cebd842669b802a4e43e91774) in parent group-v288131. [ 840.212540] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Creating folder: Instances. Parent ref: group-v288186. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 840.212814] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Waiting for the task: (returnval){ [ 840.212814] env[62914]: value = "task-1352514" [ 840.212814] env[62914]: _type = "Task" [ 840.212814] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.213283] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d899822f-a02a-432e-8f67-279c2531d114 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.225621] env[62914]: DEBUG nova.compute.provider_tree [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.232587] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352514, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.235662] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Created folder: Instances in parent group-v288186. [ 840.235662] env[62914]: DEBUG oslo.service.loopingcall [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.237040] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 840.237040] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-127e48cb-66fb-4986-b71a-05fb1fec5b90 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.258016] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 840.258016] env[62914]: value = "task-1352516" [ 840.258016] env[62914]: _type = "Task" [ 840.258016] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.265589] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352516, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.424696] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "cb89c84f-414b-4dc7-9db5-12cbc30b52b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.910s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.476869] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352511, 'name': Rename_Task, 'duration_secs': 0.153431} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.477090] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 840.477195] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-71ddc38e-07bd-46cc-9f72-ae0b71a8ee5a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.484536] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 840.484536] env[62914]: value = "task-1352517" [ 840.484536] env[62914]: _type = "Task" [ 840.484536] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.494876] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352517, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.516953] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352510, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.546624] env[62914]: DEBUG oslo_vmware.api [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352512, 'name': ResetVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.682372] env[62914]: DEBUG nova.compute.manager [req-d59624c4-44c6-4cb5-82a1-eb7c1529a6df req-8e7a9f9e-d3f3-45d1-b252-8390ff6024c9 service nova] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Received event network-vif-plugged-a5b49567-42ef-4e73-ab33-113aa689e83b {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.682602] env[62914]: DEBUG oslo_concurrency.lockutils [req-d59624c4-44c6-4cb5-82a1-eb7c1529a6df req-8e7a9f9e-d3f3-45d1-b252-8390ff6024c9 service nova] Acquiring lock "2274ca27-8e2d-435e-8570-97eb4ae3cc58-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.682931] env[62914]: DEBUG oslo_concurrency.lockutils [req-d59624c4-44c6-4cb5-82a1-eb7c1529a6df req-8e7a9f9e-d3f3-45d1-b252-8390ff6024c9 service nova] Lock "2274ca27-8e2d-435e-8570-97eb4ae3cc58-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.683140] env[62914]: DEBUG oslo_concurrency.lockutils [req-d59624c4-44c6-4cb5-82a1-eb7c1529a6df req-8e7a9f9e-d3f3-45d1-b252-8390ff6024c9 service nova] Lock "2274ca27-8e2d-435e-8570-97eb4ae3cc58-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.683327] env[62914]: DEBUG nova.compute.manager [req-d59624c4-44c6-4cb5-82a1-eb7c1529a6df req-8e7a9f9e-d3f3-45d1-b252-8390ff6024c9 service nova] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] No waiting events found dispatching network-vif-plugged-a5b49567-42ef-4e73-ab33-113aa689e83b {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 840.683507] env[62914]: WARNING nova.compute.manager [req-d59624c4-44c6-4cb5-82a1-eb7c1529a6df req-8e7a9f9e-d3f3-45d1-b252-8390ff6024c9 service nova] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Received unexpected event network-vif-plugged-a5b49567-42ef-4e73-ab33-113aa689e83b for instance with vm_state building and task_state spawning. [ 840.683694] env[62914]: DEBUG nova.compute.manager [req-d59624c4-44c6-4cb5-82a1-eb7c1529a6df req-8e7a9f9e-d3f3-45d1-b252-8390ff6024c9 service nova] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Received event network-changed-a5b49567-42ef-4e73-ab33-113aa689e83b {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.683877] env[62914]: DEBUG nova.compute.manager [req-d59624c4-44c6-4cb5-82a1-eb7c1529a6df req-8e7a9f9e-d3f3-45d1-b252-8390ff6024c9 service nova] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Refreshing instance network info cache due to event network-changed-a5b49567-42ef-4e73-ab33-113aa689e83b. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 840.684086] env[62914]: DEBUG oslo_concurrency.lockutils [req-d59624c4-44c6-4cb5-82a1-eb7c1529a6df req-8e7a9f9e-d3f3-45d1-b252-8390ff6024c9 service nova] Acquiring lock "refresh_cache-2274ca27-8e2d-435e-8570-97eb4ae3cc58" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.684250] env[62914]: DEBUG oslo_concurrency.lockutils [req-d59624c4-44c6-4cb5-82a1-eb7c1529a6df req-8e7a9f9e-d3f3-45d1-b252-8390ff6024c9 service nova] Acquired lock "refresh_cache-2274ca27-8e2d-435e-8570-97eb4ae3cc58" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.684418] env[62914]: DEBUG nova.network.neutron [req-d59624c4-44c6-4cb5-82a1-eb7c1529a6df req-8e7a9f9e-d3f3-45d1-b252-8390ff6024c9 service nova] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Refreshing network info cache for port a5b49567-42ef-4e73-ab33-113aa689e83b {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.725399] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352514, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505649} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.725713] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] bd3e82ce-62e7-4761-bcd7-d1f835234b97/bd3e82ce-62e7-4761-bcd7-d1f835234b97.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 840.725960] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.726240] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef51b5ec-bf1e-42c9-8c4e-f23cc4a36a06 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.728945] env[62914]: DEBUG nova.scheduler.client.report [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.736392] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Waiting for the task: (returnval){ [ 840.736392] env[62914]: value = "task-1352518" [ 840.736392] env[62914]: _type = "Task" [ 840.736392] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.744373] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352518, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.767016] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352516, 'name': CreateVM_Task, 'duration_secs': 0.352442} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.767207] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 840.767854] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.768026] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.768340] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 840.768588] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b30ddcb2-d6bf-40d1-9061-7ea1d92b1596 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.773056] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 840.773056] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5232dc03-6092-9296-9891-7a5ded1497a0" [ 840.773056] env[62914]: _type = "Task" [ 840.773056] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.783112] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5232dc03-6092-9296-9891-7a5ded1497a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.906310] env[62914]: DEBUG nova.compute.manager [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 840.928032] env[62914]: DEBUG nova.compute.manager [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 840.944480] env[62914]: DEBUG nova.virt.hardware [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 840.944786] env[62914]: DEBUG nova.virt.hardware [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 840.944944] env[62914]: DEBUG nova.virt.hardware [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.946613] env[62914]: DEBUG nova.virt.hardware [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 840.946613] env[62914]: DEBUG nova.virt.hardware [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.946613] env[62914]: DEBUG nova.virt.hardware [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 840.946613] env[62914]: DEBUG nova.virt.hardware [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 840.946613] env[62914]: DEBUG nova.virt.hardware [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 840.946835] env[62914]: DEBUG nova.virt.hardware [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 840.946835] env[62914]: DEBUG nova.virt.hardware [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 840.946835] env[62914]: DEBUG nova.virt.hardware [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 840.947840] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feffb227-bea2-408d-b3f5-d2f84f687315 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.958265] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-542153e0-980a-45fb-833a-123e3fcc09f4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.994978] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352517, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.014396] env[62914]: DEBUG oslo_vmware.api [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352510, 'name': PowerOnVM_Task, 'duration_secs': 1.429879} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.014727] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.014865] env[62914]: INFO nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Took 9.57 seconds to spawn the instance on the hypervisor. [ 841.015471] env[62914]: DEBUG nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.015811] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b08bb6d3-e9a7-4b54-99a9-f2ed3a7ac95f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.045647] env[62914]: DEBUG oslo_vmware.api [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352512, 'name': ResetVM_Task, 'duration_secs': 0.520185} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.045902] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Did hard reboot of VM {{(pid=62914) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 841.046227] env[62914]: DEBUG nova.compute.manager [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.047080] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb1e7b0-5634-4098-935d-5966ac0299fc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.141139] env[62914]: DEBUG nova.compute.manager [req-05767622-2447-48ea-b4fb-9445605dc33d req-48c68d54-e121-4288-b895-83b3d348a952 service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Received event network-vif-plugged-59eaafaa-3160-4236-ab11-eb6f5e3ffec8 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.141139] env[62914]: DEBUG oslo_concurrency.lockutils [req-05767622-2447-48ea-b4fb-9445605dc33d req-48c68d54-e121-4288-b895-83b3d348a952 service nova] Acquiring lock "1fe15423-f8be-4763-b55e-2cbb383ff01d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.141443] env[62914]: DEBUG oslo_concurrency.lockutils [req-05767622-2447-48ea-b4fb-9445605dc33d req-48c68d54-e121-4288-b895-83b3d348a952 service nova] Lock "1fe15423-f8be-4763-b55e-2cbb383ff01d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.141768] env[62914]: DEBUG oslo_concurrency.lockutils [req-05767622-2447-48ea-b4fb-9445605dc33d req-48c68d54-e121-4288-b895-83b3d348a952 service nova] Lock "1fe15423-f8be-4763-b55e-2cbb383ff01d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.142067] env[62914]: DEBUG nova.compute.manager [req-05767622-2447-48ea-b4fb-9445605dc33d req-48c68d54-e121-4288-b895-83b3d348a952 service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] No waiting events found dispatching network-vif-plugged-59eaafaa-3160-4236-ab11-eb6f5e3ffec8 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 841.142382] env[62914]: WARNING nova.compute.manager [req-05767622-2447-48ea-b4fb-9445605dc33d req-48c68d54-e121-4288-b895-83b3d348a952 service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Received unexpected event network-vif-plugged-59eaafaa-3160-4236-ab11-eb6f5e3ffec8 for instance with vm_state building and task_state spawning. [ 841.237242] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.360s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.237242] env[62914]: DEBUG nova.compute.manager [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 841.237926] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 11.116s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.238291] env[62914]: DEBUG nova.objects.instance [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62914) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 841.257295] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352518, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06645} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.257705] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.260459] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd81737-9047-45d2-b753-efde741aaa1b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.283775] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] bd3e82ce-62e7-4761-bcd7-d1f835234b97/bd3e82ce-62e7-4761-bcd7-d1f835234b97.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.290292] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58fb1f50-05db-4c42-be7b-27c55c4c10a0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.313516] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5232dc03-6092-9296-9891-7a5ded1497a0, 'name': SearchDatastore_Task, 'duration_secs': 0.009137} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.314951] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.315254] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 841.315547] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.315723] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.315932] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 841.316305] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Waiting for the task: (returnval){ [ 841.316305] env[62914]: value = "task-1352519" [ 841.316305] env[62914]: _type = "Task" [ 841.316305] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.316524] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-092cc806-1d29-46e7-950c-96ff68001db6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.326716] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352519, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.330880] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 841.331114] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 841.331943] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16770764-f028-49d2-a4f3-67809b84f91d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.336161] env[62914]: DEBUG nova.network.neutron [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Successfully updated port: 59eaafaa-3160-4236-ab11-eb6f5e3ffec8 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 841.339181] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 841.339181] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]522dc96c-d679-8d9f-a193-0a49c0a9d277" [ 841.339181] env[62914]: _type = "Task" [ 841.339181] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.348741] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]522dc96c-d679-8d9f-a193-0a49c0a9d277, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.456581] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.469117] env[62914]: DEBUG nova.network.neutron [req-d59624c4-44c6-4cb5-82a1-eb7c1529a6df req-8e7a9f9e-d3f3-45d1-b252-8390ff6024c9 service nova] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Updated VIF entry in instance network info cache for port a5b49567-42ef-4e73-ab33-113aa689e83b. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 841.469544] env[62914]: DEBUG nova.network.neutron [req-d59624c4-44c6-4cb5-82a1-eb7c1529a6df req-8e7a9f9e-d3f3-45d1-b252-8390ff6024c9 service nova] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Updating instance_info_cache with network_info: [{"id": "a5b49567-42ef-4e73-ab33-113aa689e83b", "address": "fa:16:3e:d8:88:64", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5b49567-42", "ovs_interfaceid": "a5b49567-42ef-4e73-ab33-113aa689e83b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.496202] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352517, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.538261] env[62914]: INFO nova.compute.manager [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Took 29.42 seconds to build instance. [ 841.570664] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c630b6cf-9197-4e72-b21a-80433a9bced0 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "46d139a1-a4ae-435e-9d6f-cfc06d706128" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.392s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.743776] env[62914]: DEBUG nova.compute.utils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 841.748257] env[62914]: DEBUG nova.compute.manager [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 841.748351] env[62914]: DEBUG nova.network.neutron [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 841.823914] env[62914]: DEBUG nova.policy [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'baf190d4a357468a9eb18dfad17e3da0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '15315f8e77d749a4b2b0211c3b22b546', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 841.831541] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352519, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.841360] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "refresh_cache-1fe15423-f8be-4763-b55e-2cbb383ff01d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.841475] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquired lock "refresh_cache-1fe15423-f8be-4763-b55e-2cbb383ff01d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.841642] env[62914]: DEBUG nova.network.neutron [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 841.852285] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]522dc96c-d679-8d9f-a193-0a49c0a9d277, 'name': SearchDatastore_Task, 'duration_secs': 0.013008} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.853126] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94f2a9e7-1f0d-4e7e-9a39-f0cd92977918 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.859214] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 841.859214] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]523dec77-4bf9-480a-d462-64f1f14999f9" [ 841.859214] env[62914]: _type = "Task" [ 841.859214] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.869050] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]523dec77-4bf9-480a-d462-64f1f14999f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.972165] env[62914]: DEBUG oslo_concurrency.lockutils [req-d59624c4-44c6-4cb5-82a1-eb7c1529a6df req-8e7a9f9e-d3f3-45d1-b252-8390ff6024c9 service nova] Releasing lock "refresh_cache-2274ca27-8e2d-435e-8570-97eb4ae3cc58" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.996528] env[62914]: DEBUG oslo_vmware.api [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352517, 'name': PowerOnVM_Task, 'duration_secs': 1.088063} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.996844] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.997060] env[62914]: INFO nova.compute.manager [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Took 8.15 seconds to spawn the instance on the hypervisor. [ 841.997362] env[62914]: DEBUG nova.compute.manager [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.998197] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05821504-3f21-4d5f-9fff-ea51ba2ea94d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.045834] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d691ef3b-790b-4e35-ad9f-cc6ec3f69730 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "9577394d-1c73-4ed1-ba86-e7c246e32719" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.506s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.209838] env[62914]: DEBUG nova.compute.manager [req-91559d07-8c74-42c6-bea5-62c3b75d99d7 req-cabbc5cb-823b-460b-9b73-3a85d3655356 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Received event network-changed-4f7e867d-96a4-4d4e-82fa-43469303566c {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.210395] env[62914]: DEBUG nova.compute.manager [req-91559d07-8c74-42c6-bea5-62c3b75d99d7 req-cabbc5cb-823b-460b-9b73-3a85d3655356 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Refreshing instance network info cache due to event network-changed-4f7e867d-96a4-4d4e-82fa-43469303566c. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 842.210665] env[62914]: DEBUG oslo_concurrency.lockutils [req-91559d07-8c74-42c6-bea5-62c3b75d99d7 req-cabbc5cb-823b-460b-9b73-3a85d3655356 service nova] Acquiring lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.210833] env[62914]: DEBUG oslo_concurrency.lockutils [req-91559d07-8c74-42c6-bea5-62c3b75d99d7 req-cabbc5cb-823b-460b-9b73-3a85d3655356 service nova] Acquired lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.210998] env[62914]: DEBUG nova.network.neutron [req-91559d07-8c74-42c6-bea5-62c3b75d99d7 req-cabbc5cb-823b-460b-9b73-3a85d3655356 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Refreshing network info cache for port 4f7e867d-96a4-4d4e-82fa-43469303566c {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 842.249437] env[62914]: DEBUG nova.compute.manager [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 842.257011] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82f4991e-374b-45a3-9c9b-a60828cc3cd1 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.258395] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.917s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.260921] env[62914]: INFO nova.compute.claims [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 842.334631] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352519, 'name': ReconfigVM_Task, 'duration_secs': 0.610952} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.334925] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Reconfigured VM instance instance-00000045 to attach disk [datastore1] bd3e82ce-62e7-4761-bcd7-d1f835234b97/bd3e82ce-62e7-4761-bcd7-d1f835234b97.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.335583] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39bc2529-b696-44b1-9dd3-ef46f8f939e1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.344981] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Waiting for the task: (returnval){ [ 842.344981] env[62914]: value = "task-1352520" [ 842.344981] env[62914]: _type = "Task" [ 842.344981] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.357643] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352520, 'name': Rename_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.371795] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]523dec77-4bf9-480a-d462-64f1f14999f9, 'name': SearchDatastore_Task, 'duration_secs': 0.010162} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.371973] env[62914]: DEBUG nova.network.neutron [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Successfully created port: 57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 842.373887] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.374183] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 2274ca27-8e2d-435e-8570-97eb4ae3cc58/2274ca27-8e2d-435e-8570-97eb4ae3cc58.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 842.374493] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c2979fd-de9c-443b-8ff7-5a1529ef19cf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.381036] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 842.381036] env[62914]: value = "task-1352521" [ 842.381036] env[62914]: _type = "Task" [ 842.381036] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.388594] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352521, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.419566] env[62914]: DEBUG nova.network.neutron [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.518534] env[62914]: INFO nova.compute.manager [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Took 28.59 seconds to build instance. [ 842.608894] env[62914]: DEBUG nova.network.neutron [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Updating instance_info_cache with network_info: [{"id": "59eaafaa-3160-4236-ab11-eb6f5e3ffec8", "address": "fa:16:3e:aa:2e:4c", "network": {"id": "184ac684-c4e6-4468-aee8-1625d96dc0b4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1513916205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3055903b6f4443bbeb9897e75e34c49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59eaafaa-31", "ovs_interfaceid": "59eaafaa-3160-4236-ab11-eb6f5e3ffec8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.827713] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "46d139a1-a4ae-435e-9d6f-cfc06d706128" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.827989] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "46d139a1-a4ae-435e-9d6f-cfc06d706128" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.828235] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "46d139a1-a4ae-435e-9d6f-cfc06d706128-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.828424] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "46d139a1-a4ae-435e-9d6f-cfc06d706128-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.829113] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "46d139a1-a4ae-435e-9d6f-cfc06d706128-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.835290] env[62914]: INFO nova.compute.manager [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Terminating instance [ 842.837878] env[62914]: DEBUG nova.compute.manager [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 842.838096] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 842.839050] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e73a3f1-f5f7-466f-a647-d80dd63967d8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.854274] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 842.857928] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 842.858179] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9983ba1f-779e-4768-95ba-8910f37892cd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.863211] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 842.868278] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352520, 'name': Rename_Task, 'duration_secs': 0.140495} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.869514] env[62914]: DEBUG nova.compute.manager [req-f1933468-7914-41b6-9624-9705504db658 req-68a1e793-c653-471d-9903-b64f40ce9d5d service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Received event network-changed-59eaafaa-3160-4236-ab11-eb6f5e3ffec8 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.869514] env[62914]: DEBUG nova.compute.manager [req-f1933468-7914-41b6-9624-9705504db658 req-68a1e793-c653-471d-9903-b64f40ce9d5d service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Refreshing instance network info cache due to event network-changed-59eaafaa-3160-4236-ab11-eb6f5e3ffec8. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 842.869514] env[62914]: DEBUG oslo_concurrency.lockutils [req-f1933468-7914-41b6-9624-9705504db658 req-68a1e793-c653-471d-9903-b64f40ce9d5d service nova] Acquiring lock "refresh_cache-1fe15423-f8be-4763-b55e-2cbb383ff01d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.870916] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.870916] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb92c14c-30d6-43ab-a9cc-c9faf2b955f9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.874708] env[62914]: DEBUG oslo_vmware.api [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Waiting for the task: (returnval){ [ 842.874708] env[62914]: value = "task-1352522" [ 842.874708] env[62914]: _type = "Task" [ 842.874708] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.888323] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Waiting for the task: (returnval){ [ 842.888323] env[62914]: value = "task-1352523" [ 842.888323] env[62914]: _type = "Task" [ 842.888323] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.912095] env[62914]: DEBUG oslo_vmware.api [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352522, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.917454] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352521, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.52006} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.924658] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 2274ca27-8e2d-435e-8570-97eb4ae3cc58/2274ca27-8e2d-435e-8570-97eb4ae3cc58.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 842.924895] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 842.925533] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352523, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.925760] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca1a51da-f061-4593-bc36-f639700ccdfe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.932531] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 842.932531] env[62914]: value = "task-1352524" [ 842.932531] env[62914]: _type = "Task" [ 842.932531] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.940940] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352524, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.022524] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373d9b7f-7802-4d85-b4ab-99ddfdedbe81 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lock "93d420a1-6d8f-4919-a42f-55aebab853ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.021s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.101749] env[62914]: DEBUG nova.network.neutron [req-91559d07-8c74-42c6-bea5-62c3b75d99d7 req-cabbc5cb-823b-460b-9b73-3a85d3655356 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Updated VIF entry in instance network info cache for port 4f7e867d-96a4-4d4e-82fa-43469303566c. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 843.102150] env[62914]: DEBUG nova.network.neutron [req-91559d07-8c74-42c6-bea5-62c3b75d99d7 req-cabbc5cb-823b-460b-9b73-3a85d3655356 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Updating instance_info_cache with network_info: [{"id": "4f7e867d-96a4-4d4e-82fa-43469303566c", "address": "fa:16:3e:51:58:7b", "network": {"id": "c8cdb39e-9415-4f67-8199-fa1b043d04d8", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-373473409-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f897e1c819f44362857de4d4398ca613", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f7e867d-96", "ovs_interfaceid": "4f7e867d-96a4-4d4e-82fa-43469303566c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.112734] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Releasing lock "refresh_cache-1fe15423-f8be-4763-b55e-2cbb383ff01d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.113054] env[62914]: DEBUG nova.compute.manager [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Instance network_info: |[{"id": "59eaafaa-3160-4236-ab11-eb6f5e3ffec8", "address": "fa:16:3e:aa:2e:4c", "network": {"id": "184ac684-c4e6-4468-aee8-1625d96dc0b4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1513916205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3055903b6f4443bbeb9897e75e34c49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59eaafaa-31", "ovs_interfaceid": "59eaafaa-3160-4236-ab11-eb6f5e3ffec8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 843.113615] env[62914]: DEBUG oslo_concurrency.lockutils [req-f1933468-7914-41b6-9624-9705504db658 req-68a1e793-c653-471d-9903-b64f40ce9d5d service nova] Acquired lock "refresh_cache-1fe15423-f8be-4763-b55e-2cbb383ff01d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.113795] env[62914]: DEBUG nova.network.neutron [req-f1933468-7914-41b6-9624-9705504db658 req-68a1e793-c653-471d-9903-b64f40ce9d5d service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Refreshing network info cache for port 59eaafaa-3160-4236-ab11-eb6f5e3ffec8 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 843.114997] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:2e:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8459aaf-d6a8-46fb-ad14-464ac3104695', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '59eaafaa-3160-4236-ab11-eb6f5e3ffec8', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.122918] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Creating folder: Project (c3055903b6f4443bbeb9897e75e34c49). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 843.124020] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6202f63f-a4a2-4ca2-be29-b43cc7df89ba {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.133891] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Created folder: Project (c3055903b6f4443bbeb9897e75e34c49) in parent group-v288131. [ 843.134107] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Creating folder: Instances. Parent ref: group-v288189. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 843.134348] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e21262df-6f24-4e4e-9021-ecf7961f542c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.143173] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Created folder: Instances in parent group-v288189. [ 843.143414] env[62914]: DEBUG oslo.service.loopingcall [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.143605] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 843.143813] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca4b3e9d-bee1-485b-b05d-f66ac7f44392 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.164582] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.164582] env[62914]: value = "task-1352527" [ 843.164582] env[62914]: _type = "Task" [ 843.164582] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.175082] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "cb89c84f-414b-4dc7-9db5-12cbc30b52b4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.175413] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "cb89c84f-414b-4dc7-9db5-12cbc30b52b4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.175687] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "cb89c84f-414b-4dc7-9db5-12cbc30b52b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.175936] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "cb89c84f-414b-4dc7-9db5-12cbc30b52b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.176190] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "cb89c84f-414b-4dc7-9db5-12cbc30b52b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.178442] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352527, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.178955] env[62914]: INFO nova.compute.manager [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Terminating instance [ 843.181498] env[62914]: DEBUG nova.compute.manager [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 843.181758] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 843.182849] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cda733ae-32d1-45df-8799-a9e6d21ba189 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.191393] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 843.192065] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dbe2c4b6-86b4-438a-89d6-0a398632646e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.199024] env[62914]: DEBUG oslo_vmware.api [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 843.199024] env[62914]: value = "task-1352528" [ 843.199024] env[62914]: _type = "Task" [ 843.199024] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.206760] env[62914]: DEBUG oslo_vmware.api [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352528, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.265300] env[62914]: DEBUG nova.compute.manager [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 843.291921] env[62914]: DEBUG nova.virt.hardware [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.292199] env[62914]: DEBUG nova.virt.hardware [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.292259] env[62914]: DEBUG nova.virt.hardware [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.292423] env[62914]: DEBUG nova.virt.hardware [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.292564] env[62914]: DEBUG nova.virt.hardware [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.292705] env[62914]: DEBUG nova.virt.hardware [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.292909] env[62914]: DEBUG nova.virt.hardware [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.293074] env[62914]: DEBUG nova.virt.hardware [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.293242] env[62914]: DEBUG nova.virt.hardware [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.293440] env[62914]: DEBUG nova.virt.hardware [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.293682] env[62914]: DEBUG nova.virt.hardware [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.294653] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8061df3-13e8-47cb-b6b7-df7b204ee4b3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.305654] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebdbc707-02cc-404a-a43c-ea48549169fd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.337831] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "9577394d-1c73-4ed1-ba86-e7c246e32719" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.338129] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "9577394d-1c73-4ed1-ba86-e7c246e32719" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.338343] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "9577394d-1c73-4ed1-ba86-e7c246e32719-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.338549] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "9577394d-1c73-4ed1-ba86-e7c246e32719-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.339354] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "9577394d-1c73-4ed1-ba86-e7c246e32719-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.341031] env[62914]: INFO nova.compute.manager [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Terminating instance [ 843.342952] env[62914]: DEBUG nova.compute.manager [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 843.343175] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 843.344103] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc5b178d-d5f2-4c5d-9d18-60c52fc4a43a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.351428] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 843.351759] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d4e24ab0-2a5b-4148-873b-f64d5d55fd62 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.362414] env[62914]: DEBUG oslo_vmware.api [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 843.362414] env[62914]: value = "task-1352529" [ 843.362414] env[62914]: _type = "Task" [ 843.362414] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.372942] env[62914]: DEBUG oslo_vmware.api [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352529, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.376157] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 843.376337] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Starting heal instance info cache {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 843.376803] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Rebuilding the list of instances to heal {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 843.387726] env[62914]: DEBUG oslo_vmware.api [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352522, 'name': PowerOffVM_Task, 'duration_secs': 0.311003} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.388011] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 843.388186] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 843.388463] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5d653f2-133e-436b-92fb-d464fa140451 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.411997] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352523, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.446967] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352524, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067052} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.449805] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 843.450997] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e856ac7-6b75-439d-a1ba-2974ee67c54a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.478911] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 2274ca27-8e2d-435e-8570-97eb4ae3cc58/2274ca27-8e2d-435e-8570-97eb4ae3cc58.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 843.479281] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 843.479511] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 843.479701] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Deleting the datastore file [datastore1] 46d139a1-a4ae-435e-9d6f-cfc06d706128 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.482986] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5498a940-cd2c-40c7-b636-20939455848a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.499120] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8b0c9d54-7931-4ebf-9925-7774eaf7485a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.507333] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 843.507333] env[62914]: value = "task-1352532" [ 843.507333] env[62914]: _type = "Task" [ 843.507333] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.508719] env[62914]: DEBUG oslo_vmware.api [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Waiting for the task: (returnval){ [ 843.508719] env[62914]: value = "task-1352531" [ 843.508719] env[62914]: _type = "Task" [ 843.508719] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.524484] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352532, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.527769] env[62914]: DEBUG oslo_vmware.api [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352531, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.597750] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7eb2182-bab7-4852-803a-a4c3ed2b4bee {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.605646] env[62914]: DEBUG oslo_concurrency.lockutils [req-91559d07-8c74-42c6-bea5-62c3b75d99d7 req-cabbc5cb-823b-460b-9b73-3a85d3655356 service nova] Releasing lock "refresh_cache-46d139a1-a4ae-435e-9d6f-cfc06d706128" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.606945] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197bbde1-bca8-4429-8d30-c4a1bf6c4f4b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.666337] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-315248c7-e206-45b4-8365-d6da5f89f48a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.689316] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6589ed-c089-4873-8bc7-db2ade25e4c9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.695450] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352527, 'name': CreateVM_Task, 'duration_secs': 0.359235} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.696817] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.698329] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.698583] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.699017] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.715102] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-640f0df5-7643-4b74-afd6-f411a6f0adce {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.718193] env[62914]: DEBUG nova.compute.provider_tree [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.730945] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 843.730945] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52910327-754c-145b-2fc0-a216c10bdae0" [ 843.730945] env[62914]: _type = "Task" [ 843.730945] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.735914] env[62914]: DEBUG oslo_vmware.api [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352528, 'name': PowerOffVM_Task, 'duration_secs': 0.262617} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.741195] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 843.741475] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 843.742156] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3362811f-ad5d-439b-b79b-8cc8a828e2d1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.752609] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52910327-754c-145b-2fc0-a216c10bdae0, 'name': SearchDatastore_Task, 'duration_secs': 0.010415} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.753394] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.753725] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.754079] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.754320] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.754603] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.754954] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6fa274f8-3055-4500-9bf4-4c2971b8b526 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.768162] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.768601] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 843.773067] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-963de0ca-606e-41f5-9ce9-8c585b1fb87f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.780706] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 843.780706] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52e64e29-669d-3248-02b1-868f8d5e5dd5" [ 843.780706] env[62914]: _type = "Task" [ 843.780706] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.791546] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52e64e29-669d-3248-02b1-868f8d5e5dd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.832288] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 843.832483] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 843.832660] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Deleting the datastore file [datastore2] cb89c84f-414b-4dc7-9db5-12cbc30b52b4 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.832923] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c66759ff-ce9e-4557-a560-22a386d63cce {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.840618] env[62914]: DEBUG oslo_vmware.api [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 843.840618] env[62914]: value = "task-1352534" [ 843.840618] env[62914]: _type = "Task" [ 843.840618] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.849195] env[62914]: DEBUG oslo_vmware.api [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352534, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.868114] env[62914]: DEBUG oslo_vmware.api [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352529, 'name': PowerOffVM_Task, 'duration_secs': 0.273165} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.868476] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 843.868588] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 843.868789] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c98c0f56-ce05-4fd9-94ae-3b0b92407331 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.886396] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Skipping network cache update for instance because it is being deleted. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 843.886544] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Skipping network cache update for instance because it is being deleted. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 843.886674] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Skipping network cache update for instance because it is being deleted. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 843.886801] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 843.886921] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 843.887055] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 843.887176] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 843.887302] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 843.908142] env[62914]: DEBUG oslo_vmware.api [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352523, 'name': PowerOnVM_Task, 'duration_secs': 0.837669} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.912023] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "refresh_cache-87ef6b06-f699-4de3-8b89-854717074406" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.912023] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquired lock "refresh_cache-87ef6b06-f699-4de3-8b89-854717074406" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.912023] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Forcefully refreshing network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 843.912023] env[62914]: DEBUG nova.objects.instance [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lazy-loading 'info_cache' on Instance uuid 87ef6b06-f699-4de3-8b89-854717074406 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.912023] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.912023] env[62914]: INFO nova.compute.manager [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Took 7.69 seconds to spawn the instance on the hypervisor. [ 843.912315] env[62914]: DEBUG nova.compute.manager [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.912315] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94dd5f8f-efd5-496f-ade7-19ed2d5dff05 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.937679] env[62914]: DEBUG nova.network.neutron [req-f1933468-7914-41b6-9624-9705504db658 req-68a1e793-c653-471d-9903-b64f40ce9d5d service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Updated VIF entry in instance network info cache for port 59eaafaa-3160-4236-ab11-eb6f5e3ffec8. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 843.938028] env[62914]: DEBUG nova.network.neutron [req-f1933468-7914-41b6-9624-9705504db658 req-68a1e793-c653-471d-9903-b64f40ce9d5d service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Updating instance_info_cache with network_info: [{"id": "59eaafaa-3160-4236-ab11-eb6f5e3ffec8", "address": "fa:16:3e:aa:2e:4c", "network": {"id": "184ac684-c4e6-4468-aee8-1625d96dc0b4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1513916205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3055903b6f4443bbeb9897e75e34c49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59eaafaa-31", "ovs_interfaceid": "59eaafaa-3160-4236-ab11-eb6f5e3ffec8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.956242] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 843.956458] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 843.956658] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Deleting the datastore file [datastore2] 9577394d-1c73-4ed1-ba86-e7c246e32719 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.957128] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be612ce0-6a20-4661-bf6c-fbd37b698428 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.963656] env[62914]: DEBUG oslo_vmware.api [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for the task: (returnval){ [ 843.963656] env[62914]: value = "task-1352536" [ 843.963656] env[62914]: _type = "Task" [ 843.963656] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.971648] env[62914]: DEBUG oslo_vmware.api [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352536, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.019292] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352532, 'name': ReconfigVM_Task, 'duration_secs': 0.298578} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.020181] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 2274ca27-8e2d-435e-8570-97eb4ae3cc58/2274ca27-8e2d-435e-8570-97eb4ae3cc58.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 844.021340] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f197e84b-2f5b-4aa9-8075-0aef8960809f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.026406] env[62914]: DEBUG oslo_vmware.api [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Task: {'id': task-1352531, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209712} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.027096] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 844.027434] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 844.027731] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 844.028029] env[62914]: INFO nova.compute.manager [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Took 1.19 seconds to destroy the instance on the hypervisor. [ 844.028376] env[62914]: DEBUG oslo.service.loopingcall [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.028659] env[62914]: DEBUG nova.compute.manager [-] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.028839] env[62914]: DEBUG nova.network.neutron [-] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 844.031630] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 844.031630] env[62914]: value = "task-1352537" [ 844.031630] env[62914]: _type = "Task" [ 844.031630] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.039916] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352537, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.152031] env[62914]: DEBUG nova.network.neutron [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Successfully updated port: 57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 844.225029] env[62914]: DEBUG nova.scheduler.client.report [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 844.250385] env[62914]: DEBUG nova.compute.manager [req-5d08c8df-38af-4d49-8b27-373b4aaa0b43 req-04465166-c20b-4559-8fd7-3db8be431cab service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Received event network-vif-plugged-57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.250385] env[62914]: DEBUG oslo_concurrency.lockutils [req-5d08c8df-38af-4d49-8b27-373b4aaa0b43 req-04465166-c20b-4559-8fd7-3db8be431cab service nova] Acquiring lock "c7cf71d7-b823-471f-949f-9ef2b539d972-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.250385] env[62914]: DEBUG oslo_concurrency.lockutils [req-5d08c8df-38af-4d49-8b27-373b4aaa0b43 req-04465166-c20b-4559-8fd7-3db8be431cab service nova] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.250385] env[62914]: DEBUG oslo_concurrency.lockutils [req-5d08c8df-38af-4d49-8b27-373b4aaa0b43 req-04465166-c20b-4559-8fd7-3db8be431cab service nova] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.250385] env[62914]: DEBUG nova.compute.manager [req-5d08c8df-38af-4d49-8b27-373b4aaa0b43 req-04465166-c20b-4559-8fd7-3db8be431cab service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] No waiting events found dispatching network-vif-plugged-57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 844.250535] env[62914]: WARNING nova.compute.manager [req-5d08c8df-38af-4d49-8b27-373b4aaa0b43 req-04465166-c20b-4559-8fd7-3db8be431cab service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Received unexpected event network-vif-plugged-57172af7-5ed6-46d9-9d7b-3ef100c530f7 for instance with vm_state building and task_state spawning. [ 844.250807] env[62914]: DEBUG nova.compute.manager [req-5d08c8df-38af-4d49-8b27-373b4aaa0b43 req-04465166-c20b-4559-8fd7-3db8be431cab service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Received event network-changed-57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.251137] env[62914]: DEBUG nova.compute.manager [req-5d08c8df-38af-4d49-8b27-373b4aaa0b43 req-04465166-c20b-4559-8fd7-3db8be431cab service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Refreshing instance network info cache due to event network-changed-57172af7-5ed6-46d9-9d7b-3ef100c530f7. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 844.251478] env[62914]: DEBUG oslo_concurrency.lockutils [req-5d08c8df-38af-4d49-8b27-373b4aaa0b43 req-04465166-c20b-4559-8fd7-3db8be431cab service nova] Acquiring lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.251740] env[62914]: DEBUG oslo_concurrency.lockutils [req-5d08c8df-38af-4d49-8b27-373b4aaa0b43 req-04465166-c20b-4559-8fd7-3db8be431cab service nova] Acquired lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.252483] env[62914]: DEBUG nova.network.neutron [req-5d08c8df-38af-4d49-8b27-373b4aaa0b43 req-04465166-c20b-4559-8fd7-3db8be431cab service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Refreshing network info cache for port 57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.291075] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52e64e29-669d-3248-02b1-868f8d5e5dd5, 'name': SearchDatastore_Task, 'duration_secs': 0.014034} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.291930] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa1d87c4-18d0-4963-a773-2cc2dbfe50e3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.297500] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 844.297500] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]524b608b-d7b1-c1a9-e917-0b39d23dc00e" [ 844.297500] env[62914]: _type = "Task" [ 844.297500] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.305760] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]524b608b-d7b1-c1a9-e917-0b39d23dc00e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.352254] env[62914]: DEBUG oslo_vmware.api [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352534, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.369218} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.352254] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 844.352254] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 844.352254] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 844.352254] env[62914]: INFO nova.compute.manager [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Took 1.17 seconds to destroy the instance on the hypervisor. [ 844.352404] env[62914]: DEBUG oslo.service.loopingcall [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.352404] env[62914]: DEBUG nova.compute.manager [-] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.352404] env[62914]: DEBUG nova.network.neutron [-] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 844.430048] env[62914]: INFO nova.compute.manager [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Took 28.61 seconds to build instance. [ 844.440182] env[62914]: DEBUG oslo_concurrency.lockutils [req-f1933468-7914-41b6-9624-9705504db658 req-68a1e793-c653-471d-9903-b64f40ce9d5d service nova] Releasing lock "refresh_cache-1fe15423-f8be-4763-b55e-2cbb383ff01d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.473730] env[62914]: DEBUG oslo_vmware.api [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Task: {'id': task-1352536, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.304745} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.473977] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 844.474252] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 844.474480] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 844.474696] env[62914]: INFO nova.compute.manager [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Took 1.13 seconds to destroy the instance on the hypervisor. [ 844.475128] env[62914]: DEBUG oslo.service.loopingcall [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.475528] env[62914]: DEBUG nova.compute.manager [-] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.475671] env[62914]: DEBUG nova.network.neutron [-] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 844.542904] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352537, 'name': Rename_Task, 'duration_secs': 0.130683} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.543550] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 844.543866] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0082af8-bbdd-42c8-a796-ed4c81a6a942 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.551092] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 844.551092] env[62914]: value = "task-1352538" [ 844.551092] env[62914]: _type = "Task" [ 844.551092] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.563156] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352538, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.655717] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.677767] env[62914]: DEBUG nova.compute.manager [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 844.679104] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce3307e-60e4-4341-8bdc-9359ae713aea {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.727701] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.469s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.728279] env[62914]: DEBUG nova.compute.manager [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 844.731764] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.582s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.733383] env[62914]: INFO nova.compute.claims [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 844.784663] env[62914]: DEBUG nova.network.neutron [req-5d08c8df-38af-4d49-8b27-373b4aaa0b43 req-04465166-c20b-4559-8fd7-3db8be431cab service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.807870] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]524b608b-d7b1-c1a9-e917-0b39d23dc00e, 'name': SearchDatastore_Task, 'duration_secs': 0.027496} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.808204] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.808513] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 1fe15423-f8be-4763-b55e-2cbb383ff01d/1fe15423-f8be-4763-b55e-2cbb383ff01d.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 844.808894] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae807454-f041-4b73-b16d-112ca79705f1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.816607] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 844.816607] env[62914]: value = "task-1352539" [ 844.816607] env[62914]: _type = "Task" [ 844.816607] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.824711] env[62914]: DEBUG nova.network.neutron [-] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.825926] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352539, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.867690] env[62914]: DEBUG nova.network.neutron [req-5d08c8df-38af-4d49-8b27-373b4aaa0b43 req-04465166-c20b-4559-8fd7-3db8be431cab service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.894893] env[62914]: DEBUG nova.compute.manager [req-c67d4694-cbe4-4e8c-bf42-8ec80cd13ce6 req-96f3a072-1d93-4d1e-af7e-ef9685f3fd88 service nova] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Received event network-vif-deleted-4f7e867d-96a4-4d4e-82fa-43469303566c {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.895092] env[62914]: DEBUG nova.compute.manager [req-c67d4694-cbe4-4e8c-bf42-8ec80cd13ce6 req-96f3a072-1d93-4d1e-af7e-ef9685f3fd88 service nova] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Received event network-vif-deleted-1c71fa9c-a921-4649-81d3-9de5b41619c6 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.895377] env[62914]: INFO nova.compute.manager [req-c67d4694-cbe4-4e8c-bf42-8ec80cd13ce6 req-96f3a072-1d93-4d1e-af7e-ef9685f3fd88 service nova] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Neutron deleted interface 1c71fa9c-a921-4649-81d3-9de5b41619c6; detaching it from the instance and deleting it from the info cache [ 844.895638] env[62914]: DEBUG nova.network.neutron [req-c67d4694-cbe4-4e8c-bf42-8ec80cd13ce6 req-96f3a072-1d93-4d1e-af7e-ef9685f3fd88 service nova] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.933399] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7a31880e-eb6f-419e-8d6c-cf2cc2145f12 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Lock "bd3e82ce-62e7-4761-bcd7-d1f835234b97" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.446s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.935651] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 845.063176] env[62914]: DEBUG oslo_vmware.api [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352538, 'name': PowerOnVM_Task, 'duration_secs': 0.48197} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.063470] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 845.063698] env[62914]: INFO nova.compute.manager [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Took 6.49 seconds to spawn the instance on the hypervisor. [ 845.063874] env[62914]: DEBUG nova.compute.manager [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 845.064661] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d51937-a41c-41be-bd2b-6447c1860963 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.129026] env[62914]: DEBUG nova.network.neutron [-] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.190434] env[62914]: INFO nova.compute.manager [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] instance snapshotting [ 845.193267] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475179c9-3353-43e9-86c4-1c7a5abefdb9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.213871] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88bed5c7-6fe2-465c-b7ad-38ed7618d08b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.237909] env[62914]: DEBUG nova.compute.utils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 845.241128] env[62914]: DEBUG nova.compute.manager [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 845.241297] env[62914]: DEBUG nova.network.neutron [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 845.268019] env[62914]: DEBUG nova.network.neutron [-] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.295807] env[62914]: DEBUG nova.policy [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce06463e715b4975bc43b14be64ef9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '619be7b3bf1445b68ba7adefc98f8782', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 845.326963] env[62914]: INFO nova.compute.manager [-] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Took 1.30 seconds to deallocate network for instance. [ 845.327311] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352539, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.370878] env[62914]: DEBUG oslo_concurrency.lockutils [req-5d08c8df-38af-4d49-8b27-373b4aaa0b43 req-04465166-c20b-4559-8fd7-3db8be431cab service nova] Releasing lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.371610] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquired lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.371812] env[62914]: DEBUG nova.network.neutron [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 845.399451] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-619d2069-405b-4dce-a93b-7a04fd9ddb6e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.409917] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3511b81-db88-4f23-bc25-9f2750d2140f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.449820] env[62914]: DEBUG nova.compute.manager [req-c67d4694-cbe4-4e8c-bf42-8ec80cd13ce6 req-96f3a072-1d93-4d1e-af7e-ef9685f3fd88 service nova] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Detach interface failed, port_id=1c71fa9c-a921-4649-81d3-9de5b41619c6, reason: Instance cb89c84f-414b-4dc7-9db5-12cbc30b52b4 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 845.537501] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.590759] env[62914]: INFO nova.compute.manager [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Took 23.26 seconds to build instance. [ 845.632270] env[62914]: INFO nova.compute.manager [-] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Took 1.28 seconds to deallocate network for instance. [ 845.676144] env[62914]: DEBUG nova.network.neutron [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Successfully created port: 189c8622-547d-4da9-94da-564df3bc382d {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 845.724548] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Creating Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 845.726652] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bbbe2116-4907-4558-8460-5336ed3a7b5b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.732936] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 845.732936] env[62914]: value = "task-1352540" [ 845.732936] env[62914]: _type = "Task" [ 845.732936] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.740659] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352540, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.746978] env[62914]: DEBUG nova.compute.manager [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 845.773145] env[62914]: INFO nova.compute.manager [-] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Took 1.30 seconds to deallocate network for instance. [ 845.833999] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.834392] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352539, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.807029} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.834966] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 1fe15423-f8be-4763-b55e-2cbb383ff01d/1fe15423-f8be-4763-b55e-2cbb383ff01d.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 845.835195] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.835447] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e0e4b737-3b69-4e02-84f7-a4023fd81f85 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.843470] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 845.843470] env[62914]: value = "task-1352541" [ 845.843470] env[62914]: _type = "Task" [ 845.843470] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.855970] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352541, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.919207] env[62914]: DEBUG nova.network.neutron [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.006881] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbf5261-c0ef-4978-b74c-26374c05fb1d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.015635] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac2e42f-2fc3-4d6e-af55-e2022f3af54b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.050578] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Releasing lock "refresh_cache-87ef6b06-f699-4de3-8b89-854717074406" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.050889] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Updated the network info_cache for instance {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 846.054014] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 846.055166] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0fecee-5944-4de7-82f9-2a15be8e66d0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.057992] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 846.059308] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 846.059308] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 846.059587] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 846.059695] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 846.059865] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62914) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 846.060100] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager.update_available_resource {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 846.064852] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c65f49a-3a4c-48bb-96b2-27ef2ab17579 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.079030] env[62914]: DEBUG nova.compute.provider_tree [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.094427] env[62914]: DEBUG oslo_concurrency.lockutils [None req-79180f28-e678-4ac8-b394-36332f29785a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "2274ca27-8e2d-435e-8570-97eb4ae3cc58" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.468s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.124317] env[62914]: DEBUG nova.network.neutron [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updating instance_info_cache with network_info: [{"id": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "address": "fa:16:3e:91:13:35", "network": {"id": "c4298e11-ab16-4b37-be29-48fc406e74b0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-695272838-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "15315f8e77d749a4b2b0211c3b22b546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe38bb7e-8bcb-419d-868f-0dc105c69651", "external-id": "nsx-vlan-transportzone-432", "segmentation_id": 432, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57172af7-5e", "ovs_interfaceid": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.138575] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.160171] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Acquiring lock "bd3e82ce-62e7-4761-bcd7-d1f835234b97" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.161034] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Lock "bd3e82ce-62e7-4761-bcd7-d1f835234b97" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.161270] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Acquiring lock "bd3e82ce-62e7-4761-bcd7-d1f835234b97-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.161459] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Lock "bd3e82ce-62e7-4761-bcd7-d1f835234b97-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.164403] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Lock "bd3e82ce-62e7-4761-bcd7-d1f835234b97-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.165625] env[62914]: INFO nova.compute.manager [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Terminating instance [ 846.168619] env[62914]: DEBUG nova.compute.manager [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 846.168834] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 846.169998] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb1db51-5281-4640-8601-42650ec20003 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.180685] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 846.180942] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b0cee4e-3bb5-4b52-ad3f-aaaa37227d19 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.188032] env[62914]: DEBUG oslo_vmware.api [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Waiting for the task: (returnval){ [ 846.188032] env[62914]: value = "task-1352542" [ 846.188032] env[62914]: _type = "Task" [ 846.188032] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.198075] env[62914]: DEBUG oslo_vmware.api [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352542, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.242337] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352540, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.282622] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.295475] env[62914]: DEBUG nova.compute.manager [req-e8ee4774-ed32-442c-bf0b-403b70a57195 req-5f04e77f-4277-4a12-85d1-601190ad89dd service nova] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Received event network-vif-deleted-68eeb848-a0dd-455b-8501-a7b14519e427 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.359837] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352541, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.164121} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.360429] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.361311] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f93aa7-c6e7-440b-a7c1-0e6ab581e4b4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.386044] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 1fe15423-f8be-4763-b55e-2cbb383ff01d/1fe15423-f8be-4763-b55e-2cbb383ff01d.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.386404] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fa07c63-812e-4f6f-93cb-77b232920b52 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.408039] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 846.408039] env[62914]: value = "task-1352543" [ 846.408039] env[62914]: _type = "Task" [ 846.408039] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.415611] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352543, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.565406] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.582631] env[62914]: DEBUG nova.scheduler.client.report [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.626960] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Releasing lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.627424] env[62914]: DEBUG nova.compute.manager [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Instance network_info: |[{"id": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "address": "fa:16:3e:91:13:35", "network": {"id": "c4298e11-ab16-4b37-be29-48fc406e74b0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-695272838-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "15315f8e77d749a4b2b0211c3b22b546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe38bb7e-8bcb-419d-868f-0dc105c69651", "external-id": "nsx-vlan-transportzone-432", "segmentation_id": 432, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57172af7-5e", "ovs_interfaceid": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 846.627920] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:13:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe38bb7e-8bcb-419d-868f-0dc105c69651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '57172af7-5ed6-46d9-9d7b-3ef100c530f7', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 846.636173] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Creating folder: Project (15315f8e77d749a4b2b0211c3b22b546). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 846.636728] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dfa45b37-8b33-4f6d-9374-ddfd255a081a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.647467] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Created folder: Project (15315f8e77d749a4b2b0211c3b22b546) in parent group-v288131. [ 846.647670] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Creating folder: Instances. Parent ref: group-v288192. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 846.647877] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ad4412c-34d3-4709-ad7d-feedadd3c06f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.656728] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Created folder: Instances in parent group-v288192. [ 846.656954] env[62914]: DEBUG oslo.service.loopingcall [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.657159] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 846.657678] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-196e84eb-15c7-4d8e-91f0-27fb554d95ad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.672063] env[62914]: DEBUG oslo_concurrency.lockutils [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "2274ca27-8e2d-435e-8570-97eb4ae3cc58" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.672294] env[62914]: DEBUG oslo_concurrency.lockutils [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "2274ca27-8e2d-435e-8570-97eb4ae3cc58" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.672505] env[62914]: DEBUG oslo_concurrency.lockutils [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "2274ca27-8e2d-435e-8570-97eb4ae3cc58-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.672687] env[62914]: DEBUG oslo_concurrency.lockutils [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "2274ca27-8e2d-435e-8570-97eb4ae3cc58-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.672851] env[62914]: DEBUG oslo_concurrency.lockutils [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "2274ca27-8e2d-435e-8570-97eb4ae3cc58-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.675163] env[62914]: INFO nova.compute.manager [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Terminating instance [ 846.677097] env[62914]: DEBUG nova.compute.manager [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 846.677308] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 846.678114] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e912227a-0670-4003-a6f5-dd938950dd51 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.681577] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 846.681577] env[62914]: value = "task-1352546" [ 846.681577] env[62914]: _type = "Task" [ 846.681577] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.686516] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 846.686994] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d9824b3b-fbda-43d5-9d9b-1f0de670ed95 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.690925] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352546, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.698852] env[62914]: DEBUG oslo_vmware.api [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352542, 'name': PowerOffVM_Task, 'duration_secs': 0.381588} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.699807] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 846.699984] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 846.700274] env[62914]: DEBUG oslo_vmware.api [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 846.700274] env[62914]: value = "task-1352547" [ 846.700274] env[62914]: _type = "Task" [ 846.700274] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.700451] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d576de27-a946-4864-b7c8-aa185871770b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.711884] env[62914]: DEBUG oslo_vmware.api [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352547, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.742061] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352540, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.759624] env[62914]: DEBUG nova.compute.manager [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 846.779676] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 846.779829] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 846.780043] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Deleting the datastore file [datastore1] bd3e82ce-62e7-4761-bcd7-d1f835234b97 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 846.782614] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ab7b09b-4397-4f6f-846a-3734af558363 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.789580] env[62914]: DEBUG oslo_vmware.api [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Waiting for the task: (returnval){ [ 846.789580] env[62914]: value = "task-1352549" [ 846.789580] env[62914]: _type = "Task" [ 846.789580] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.792296] env[62914]: DEBUG nova.virt.hardware [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 846.792489] env[62914]: DEBUG nova.virt.hardware [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 846.792681] env[62914]: DEBUG nova.virt.hardware [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 846.792874] env[62914]: DEBUG nova.virt.hardware [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 846.793033] env[62914]: DEBUG nova.virt.hardware [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 846.793195] env[62914]: DEBUG nova.virt.hardware [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 846.793436] env[62914]: DEBUG nova.virt.hardware [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 846.793638] env[62914]: DEBUG nova.virt.hardware [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 846.793811] env[62914]: DEBUG nova.virt.hardware [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 846.793988] env[62914]: DEBUG nova.virt.hardware [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 846.794318] env[62914]: DEBUG nova.virt.hardware [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 846.795204] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be7f839b-ab21-4661-9e73-259978bb7648 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.809314] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c870e5b-ccbe-40f3-816d-8ba4226c7738 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.813354] env[62914]: DEBUG oslo_vmware.api [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352549, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.917720] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352543, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.089564] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.090104] env[62914]: DEBUG nova.compute.manager [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 847.092739] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.528s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.092961] env[62914]: DEBUG nova.objects.instance [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Lazy-loading 'resources' on Instance uuid 87ef6b06-f699-4de3-8b89-854717074406 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.194497] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352546, 'name': CreateVM_Task, 'duration_secs': 0.337523} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.194676] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 847.195435] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.195596] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.195906] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 847.196172] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac9a2ce3-4746-4ee7-830e-564018608360 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.200972] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 847.200972] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ec047a-d2ad-44fc-c044-4b79c24e6692" [ 847.200972] env[62914]: _type = "Task" [ 847.200972] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.209109] env[62914]: DEBUG nova.compute.manager [req-842dafd9-c223-4d78-9986-bb1c9dfe257d req-ce32b055-c011-4762-ae8c-47df112e5257 service nova] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Received event network-vif-plugged-189c8622-547d-4da9-94da-564df3bc382d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.209344] env[62914]: DEBUG oslo_concurrency.lockutils [req-842dafd9-c223-4d78-9986-bb1c9dfe257d req-ce32b055-c011-4762-ae8c-47df112e5257 service nova] Acquiring lock "4bf76ce0-b5d2-4184-888c-d0ef39878356-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.209792] env[62914]: DEBUG oslo_concurrency.lockutils [req-842dafd9-c223-4d78-9986-bb1c9dfe257d req-ce32b055-c011-4762-ae8c-47df112e5257 service nova] Lock "4bf76ce0-b5d2-4184-888c-d0ef39878356-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.209792] env[62914]: DEBUG oslo_concurrency.lockutils [req-842dafd9-c223-4d78-9986-bb1c9dfe257d req-ce32b055-c011-4762-ae8c-47df112e5257 service nova] Lock "4bf76ce0-b5d2-4184-888c-d0ef39878356-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.209878] env[62914]: DEBUG nova.compute.manager [req-842dafd9-c223-4d78-9986-bb1c9dfe257d req-ce32b055-c011-4762-ae8c-47df112e5257 service nova] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] No waiting events found dispatching network-vif-plugged-189c8622-547d-4da9-94da-564df3bc382d {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 847.210056] env[62914]: WARNING nova.compute.manager [req-842dafd9-c223-4d78-9986-bb1c9dfe257d req-ce32b055-c011-4762-ae8c-47df112e5257 service nova] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Received unexpected event network-vif-plugged-189c8622-547d-4da9-94da-564df3bc382d for instance with vm_state building and task_state spawning. [ 847.213950] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ec047a-d2ad-44fc-c044-4b79c24e6692, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.217913] env[62914]: DEBUG oslo_vmware.api [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352547, 'name': PowerOffVM_Task, 'duration_secs': 0.200584} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.218159] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 847.218322] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 847.218553] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d515798f-ff00-4768-b50b-fd16b50e6020 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.241907] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352540, 'name': CreateSnapshot_Task, 'duration_secs': 1.358499} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.242165] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Created Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 847.242880] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96482f0e-7698-45b6-a9de-1080330c1859 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.280871] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 847.281085] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 847.281264] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleting the datastore file [datastore1] 2274ca27-8e2d-435e-8570-97eb4ae3cc58 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 847.281684] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-08ddb77a-535b-473a-811a-1179a518c77d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.288542] env[62914]: DEBUG oslo_vmware.api [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 847.288542] env[62914]: value = "task-1352551" [ 847.288542] env[62914]: _type = "Task" [ 847.288542] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.297084] env[62914]: DEBUG oslo_vmware.api [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352551, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.304775] env[62914]: DEBUG oslo_vmware.api [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Task: {'id': task-1352549, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161116} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.304775] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 847.304775] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 847.304917] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 847.305100] env[62914]: INFO nova.compute.manager [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Took 1.14 seconds to destroy the instance on the hypervisor. [ 847.305329] env[62914]: DEBUG oslo.service.loopingcall [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.305513] env[62914]: DEBUG nova.compute.manager [-] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 847.305604] env[62914]: DEBUG nova.network.neutron [-] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 847.371234] env[62914]: DEBUG nova.network.neutron [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Successfully updated port: 189c8622-547d-4da9-94da-564df3bc382d {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 847.417592] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352543, 'name': ReconfigVM_Task, 'duration_secs': 0.55154} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.417837] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 1fe15423-f8be-4763-b55e-2cbb383ff01d/1fe15423-f8be-4763-b55e-2cbb383ff01d.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.418473] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79115490-182d-4528-b75f-ab2ebe3ae546 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.424648] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 847.424648] env[62914]: value = "task-1352552" [ 847.424648] env[62914]: _type = "Task" [ 847.424648] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.432384] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352552, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.595991] env[62914]: DEBUG nova.compute.utils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 847.600790] env[62914]: DEBUG nova.compute.manager [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 847.601029] env[62914]: DEBUG nova.network.neutron [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 847.645546] env[62914]: DEBUG nova.policy [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41c19b40819c49d8a9dacdb74ad5899b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f5fa79f0260e461c952301f904101f79', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 847.713431] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ec047a-d2ad-44fc-c044-4b79c24e6692, 'name': SearchDatastore_Task, 'duration_secs': 0.014} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.713737] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.713974] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 847.714789] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.714789] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.714789] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 847.714789] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-31301953-e038-4370-8f03-852a1a0e139b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.725937] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 847.726374] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 847.727117] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40998330-d63a-4684-a2fe-3768926ed541 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.737277] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 847.737277] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52712881-3b3c-d460-ea6d-93d0974136db" [ 847.737277] env[62914]: _type = "Task" [ 847.737277] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.747661] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52712881-3b3c-d460-ea6d-93d0974136db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.770170] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Creating linked-clone VM from snapshot {{(pid=62914) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 847.771210] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8ece2dbb-9afc-48e8-9a3b-9dc59290705a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.786017] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 847.786017] env[62914]: value = "task-1352553" [ 847.786017] env[62914]: _type = "Task" [ 847.786017] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.803725] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352553, 'name': CloneVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.807832] env[62914]: DEBUG oslo_vmware.api [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352551, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.313375} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.808074] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 847.808257] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 847.808609] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 847.808811] env[62914]: INFO nova.compute.manager [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Took 1.13 seconds to destroy the instance on the hypervisor. [ 847.809056] env[62914]: DEBUG oslo.service.loopingcall [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.811336] env[62914]: DEBUG nova.compute.manager [-] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 847.811435] env[62914]: DEBUG nova.network.neutron [-] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 847.841984] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-380eb949-7d5e-450a-9d75-89d7804dbfac {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.850512] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-901828a8-02a3-489a-a66e-14c005ff954e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.883585] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.883585] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.883585] env[62914]: DEBUG nova.network.neutron [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 847.887021] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d841e48-f2ff-47d5-80a3-93afa905e9ca {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.893081] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7be580-116e-4c01-8f9e-f46fb402b1ea {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.909159] env[62914]: DEBUG nova.compute.provider_tree [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.909695] env[62914]: DEBUG nova.network.neutron [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Successfully created port: cf125424-b194-49ac-886d-a8b053ae7a77 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.934308] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352552, 'name': Rename_Task, 'duration_secs': 0.138196} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.934578] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 847.934816] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66d5a044-95c4-468c-a520-3bb6f628b4fe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.940656] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 847.940656] env[62914]: value = "task-1352554" [ 847.940656] env[62914]: _type = "Task" [ 847.940656] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.948413] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352554, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.095306] env[62914]: DEBUG nova.network.neutron [-] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.098482] env[62914]: DEBUG nova.compute.manager [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 848.252564] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52712881-3b3c-d460-ea6d-93d0974136db, 'name': SearchDatastore_Task, 'duration_secs': 0.01827} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.253411] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1c86541-110a-42de-b693-6ab92be48b59 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.259442] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 848.259442] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c0b44e-f9b0-0562-cd9c-9ef32d77cb04" [ 848.259442] env[62914]: _type = "Task" [ 848.259442] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.268514] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c0b44e-f9b0-0562-cd9c-9ef32d77cb04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.296221] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352553, 'name': CloneVM_Task} progress is 94%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.318740] env[62914]: DEBUG nova.compute.manager [req-e5d99cad-d5f9-40c3-8880-2e59103b54e7 req-3b86e043-b05c-442e-bf39-3b784fb50348 service nova] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Received event network-vif-deleted-0f287a59-bbed-4541-988e-f8481710291e {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.413470] env[62914]: DEBUG nova.scheduler.client.report [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.434599] env[62914]: DEBUG nova.network.neutron [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 848.451367] env[62914]: DEBUG oslo_vmware.api [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352554, 'name': PowerOnVM_Task, 'duration_secs': 0.440859} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.451367] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 848.451545] env[62914]: INFO nova.compute.manager [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Took 7.55 seconds to spawn the instance on the hypervisor. [ 848.451711] env[62914]: DEBUG nova.compute.manager [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.452480] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf50e61-162e-4a03-9eab-f506edbdf96a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.545945] env[62914]: DEBUG nova.network.neutron [-] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.598135] env[62914]: INFO nova.compute.manager [-] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Took 1.29 seconds to deallocate network for instance. [ 848.640597] env[62914]: DEBUG nova.network.neutron [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating instance_info_cache with network_info: [{"id": "189c8622-547d-4da9-94da-564df3bc382d", "address": "fa:16:3e:b6:1a:f0", "network": {"id": "21b6ff73-feab-478c-848a-313ea0acbfc2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1273337660-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "619be7b3bf1445b68ba7adefc98f8782", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap189c8622-54", "ovs_interfaceid": "189c8622-547d-4da9-94da-564df3bc382d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.770190] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c0b44e-f9b0-0562-cd9c-9ef32d77cb04, 'name': SearchDatastore_Task, 'duration_secs': 0.009911} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.770497] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.770716] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] c7cf71d7-b823-471f-949f-9ef2b539d972/c7cf71d7-b823-471f-949f-9ef2b539d972.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 848.770973] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8bd4fe3-79d6-409d-8a16-327eca9aa15f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.776715] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 848.776715] env[62914]: value = "task-1352555" [ 848.776715] env[62914]: _type = "Task" [ 848.776715] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.784227] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352555, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.796133] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352553, 'name': CloneVM_Task} progress is 94%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.922877] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.830s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.925493] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.063s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.927045] env[62914]: INFO nova.compute.claims [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.959350] env[62914]: INFO nova.scheduler.client.report [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Deleted allocations for instance 87ef6b06-f699-4de3-8b89-854717074406 [ 848.973873] env[62914]: INFO nova.compute.manager [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Took 21.93 seconds to build instance. [ 849.048299] env[62914]: INFO nova.compute.manager [-] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Took 1.24 seconds to deallocate network for instance. [ 849.112908] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.116395] env[62914]: DEBUG nova.compute.manager [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 849.141407] env[62914]: DEBUG nova.virt.hardware [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 849.141687] env[62914]: DEBUG nova.virt.hardware [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 849.141820] env[62914]: DEBUG nova.virt.hardware [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.141998] env[62914]: DEBUG nova.virt.hardware [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 849.142159] env[62914]: DEBUG nova.virt.hardware [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.142305] env[62914]: DEBUG nova.virt.hardware [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 849.142516] env[62914]: DEBUG nova.virt.hardware [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 849.142673] env[62914]: DEBUG nova.virt.hardware [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 849.142836] env[62914]: DEBUG nova.virt.hardware [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 849.142991] env[62914]: DEBUG nova.virt.hardware [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 849.143178] env[62914]: DEBUG nova.virt.hardware [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.143767] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.144091] env[62914]: DEBUG nova.compute.manager [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Instance network_info: |[{"id": "189c8622-547d-4da9-94da-564df3bc382d", "address": "fa:16:3e:b6:1a:f0", "network": {"id": "21b6ff73-feab-478c-848a-313ea0acbfc2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1273337660-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "619be7b3bf1445b68ba7adefc98f8782", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap189c8622-54", "ovs_interfaceid": "189c8622-547d-4da9-94da-564df3bc382d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 849.144939] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a0de19e-53c3-4a38-ab9a-7e9bed552794 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.147950] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:1a:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10b81051-1eb1-406b-888c-4548c470c77e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '189c8622-547d-4da9-94da-564df3bc382d', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.155627] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Creating folder: Project (619be7b3bf1445b68ba7adefc98f8782). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.155950] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ea6f6704-61ad-4585-be15-1aaa309f81b8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.163756] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e1dac21-5ddb-4ffb-a110-9a46890bdd16 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.169265] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Created folder: Project (619be7b3bf1445b68ba7adefc98f8782) in parent group-v288131. [ 849.169560] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Creating folder: Instances. Parent ref: group-v288197. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.170204] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f71aea5e-e349-47fb-8175-eb0d91ad21df {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.189406] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Created folder: Instances in parent group-v288197. [ 849.189585] env[62914]: DEBUG oslo.service.loopingcall [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.189813] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.190058] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5428cb5c-c964-4b2b-b226-574ae23412b7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.209939] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.209939] env[62914]: value = "task-1352558" [ 849.209939] env[62914]: _type = "Task" [ 849.209939] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.220755] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352558, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.238304] env[62914]: DEBUG nova.compute.manager [req-8adc8037-623f-46e8-b6d0-f2050ea8413f req-fb11139b-b5e4-4994-a378-2ef23ad6713c service nova] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Received event network-changed-189c8622-547d-4da9-94da-564df3bc382d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.238304] env[62914]: DEBUG nova.compute.manager [req-8adc8037-623f-46e8-b6d0-f2050ea8413f req-fb11139b-b5e4-4994-a378-2ef23ad6713c service nova] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Refreshing instance network info cache due to event network-changed-189c8622-547d-4da9-94da-564df3bc382d. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 849.238627] env[62914]: DEBUG oslo_concurrency.lockutils [req-8adc8037-623f-46e8-b6d0-f2050ea8413f req-fb11139b-b5e4-4994-a378-2ef23ad6713c service nova] Acquiring lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.238818] env[62914]: DEBUG oslo_concurrency.lockutils [req-8adc8037-623f-46e8-b6d0-f2050ea8413f req-fb11139b-b5e4-4994-a378-2ef23ad6713c service nova] Acquired lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.239520] env[62914]: DEBUG nova.network.neutron [req-8adc8037-623f-46e8-b6d0-f2050ea8413f req-fb11139b-b5e4-4994-a378-2ef23ad6713c service nova] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Refreshing network info cache for port 189c8622-547d-4da9-94da-564df3bc382d {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 849.291338] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352555, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.300935] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352553, 'name': CloneVM_Task} progress is 95%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.468120] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb990e3e-7f09-4160-a135-48bb83c90a12 tempest-ServerShowV254Test-1217992908 tempest-ServerShowV254Test-1217992908-project-member] Lock "87ef6b06-f699-4de3-8b89-854717074406" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.669s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.476411] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cf9bebe4-a027-4310-aa94-d2923cdd54b7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.165s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.523097] env[62914]: DEBUG nova.network.neutron [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Successfully updated port: cf125424-b194-49ac-886d-a8b053ae7a77 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.554754] env[62914]: DEBUG oslo_concurrency.lockutils [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.720692] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352558, 'name': CreateVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.787522] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352555, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.565452} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.787818] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] c7cf71d7-b823-471f-949f-9ef2b539d972/c7cf71d7-b823-471f-949f-9ef2b539d972.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 849.787883] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.788435] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-370eabac-ce4d-4901-a373-d8bb952de5bf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.796619] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 849.796619] env[62914]: value = "task-1352559" [ 849.796619] env[62914]: _type = "Task" [ 849.796619] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.806682] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352553, 'name': CloneVM_Task, 'duration_secs': 1.70146} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.810549] env[62914]: INFO nova.virt.vmwareapi.vmops [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Created linked-clone VM from snapshot [ 849.811422] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b96efd89-e63c-407e-8e14-cbdc04c6e5cc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.818202] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352559, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.823550] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Uploading image 1040c236-a1c7-4558-88d1-f521d06e9843 {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 849.856183] env[62914]: DEBUG oslo_vmware.rw_handles [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 849.856183] env[62914]: value = "vm-288196" [ 849.856183] env[62914]: _type = "VirtualMachine" [ 849.856183] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 849.856702] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-be741afd-b2bf-4a4f-bc4a-3fdc50f94241 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.864922] env[62914]: DEBUG oslo_vmware.rw_handles [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lease: (returnval){ [ 849.864922] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5259b650-5c52-7fc3-0bf3-6f49a39005ce" [ 849.864922] env[62914]: _type = "HttpNfcLease" [ 849.864922] env[62914]: } obtained for exporting VM: (result){ [ 849.864922] env[62914]: value = "vm-288196" [ 849.864922] env[62914]: _type = "VirtualMachine" [ 849.864922] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 849.865563] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the lease: (returnval){ [ 849.865563] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5259b650-5c52-7fc3-0bf3-6f49a39005ce" [ 849.865563] env[62914]: _type = "HttpNfcLease" [ 849.865563] env[62914]: } to be ready. {{(pid=62914) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 849.873185] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 849.873185] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5259b650-5c52-7fc3-0bf3-6f49a39005ce" [ 849.873185] env[62914]: _type = "HttpNfcLease" [ 849.873185] env[62914]: } is initializing. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 850.025550] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "refresh_cache-43ee8948-8805-4d48-bd45-e93d2e2eb05d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.025700] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "refresh_cache-43ee8948-8805-4d48-bd45-e93d2e2eb05d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.025844] env[62914]: DEBUG nova.network.neutron [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 850.179522] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a6b481-ca2c-48a9-8b9d-1deff9bd6eb4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.187643] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac77a0d-3636-44c0-9aed-6d34ab0178b7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.220319] env[62914]: DEBUG nova.network.neutron [req-8adc8037-623f-46e8-b6d0-f2050ea8413f req-fb11139b-b5e4-4994-a378-2ef23ad6713c service nova] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updated VIF entry in instance network info cache for port 189c8622-547d-4da9-94da-564df3bc382d. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 850.220742] env[62914]: DEBUG nova.network.neutron [req-8adc8037-623f-46e8-b6d0-f2050ea8413f req-fb11139b-b5e4-4994-a378-2ef23ad6713c service nova] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating instance_info_cache with network_info: [{"id": "189c8622-547d-4da9-94da-564df3bc382d", "address": "fa:16:3e:b6:1a:f0", "network": {"id": "21b6ff73-feab-478c-848a-313ea0acbfc2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1273337660-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "619be7b3bf1445b68ba7adefc98f8782", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap189c8622-54", "ovs_interfaceid": "189c8622-547d-4da9-94da-564df3bc382d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.225600] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f6f062-233b-4e76-aee5-f6e85d734313 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.236388] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352558, 'name': CreateVM_Task, 'duration_secs': 0.520046} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.237226] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 850.238417] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7ffd5c-0da4-4a0b-8ec2-b9484201b5ba {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.242597] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.243452] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.243571] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 850.244188] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbf8033b-a966-422c-a6a3-c749fb7d12f5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.259019] env[62914]: DEBUG nova.compute.provider_tree [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.259616] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 850.259616] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5208e8d9-2b67-44ed-04f3-106ae14e67b5" [ 850.259616] env[62914]: _type = "Task" [ 850.259616] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.267887] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5208e8d9-2b67-44ed-04f3-106ae14e67b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.307504] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352559, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107952} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.307835] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 850.308628] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-390f0ffd-7972-4e4c-a1ff-f83d73488a60 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.330026] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] c7cf71d7-b823-471f-949f-9ef2b539d972/c7cf71d7-b823-471f-949f-9ef2b539d972.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.330423] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e18bd1b5-ab37-48e9-94fe-a686449ba801 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.348128] env[62914]: DEBUG nova.compute.manager [req-2a292e54-485e-4ec9-a08a-ff343118ce28 req-f9a5810a-07ff-4904-980c-0c85d858cd35 service nova] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Received event network-vif-plugged-cf125424-b194-49ac-886d-a8b053ae7a77 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.348340] env[62914]: DEBUG oslo_concurrency.lockutils [req-2a292e54-485e-4ec9-a08a-ff343118ce28 req-f9a5810a-07ff-4904-980c-0c85d858cd35 service nova] Acquiring lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.348547] env[62914]: DEBUG oslo_concurrency.lockutils [req-2a292e54-485e-4ec9-a08a-ff343118ce28 req-f9a5810a-07ff-4904-980c-0c85d858cd35 service nova] Lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.348748] env[62914]: DEBUG oslo_concurrency.lockutils [req-2a292e54-485e-4ec9-a08a-ff343118ce28 req-f9a5810a-07ff-4904-980c-0c85d858cd35 service nova] Lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.348928] env[62914]: DEBUG nova.compute.manager [req-2a292e54-485e-4ec9-a08a-ff343118ce28 req-f9a5810a-07ff-4904-980c-0c85d858cd35 service nova] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] No waiting events found dispatching network-vif-plugged-cf125424-b194-49ac-886d-a8b053ae7a77 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 850.349097] env[62914]: WARNING nova.compute.manager [req-2a292e54-485e-4ec9-a08a-ff343118ce28 req-f9a5810a-07ff-4904-980c-0c85d858cd35 service nova] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Received unexpected event network-vif-plugged-cf125424-b194-49ac-886d-a8b053ae7a77 for instance with vm_state building and task_state spawning. [ 850.349253] env[62914]: DEBUG nova.compute.manager [req-2a292e54-485e-4ec9-a08a-ff343118ce28 req-f9a5810a-07ff-4904-980c-0c85d858cd35 service nova] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Received event network-changed-cf125424-b194-49ac-886d-a8b053ae7a77 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.349614] env[62914]: DEBUG nova.compute.manager [req-2a292e54-485e-4ec9-a08a-ff343118ce28 req-f9a5810a-07ff-4904-980c-0c85d858cd35 service nova] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Refreshing instance network info cache due to event network-changed-cf125424-b194-49ac-886d-a8b053ae7a77. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 850.349804] env[62914]: DEBUG oslo_concurrency.lockutils [req-2a292e54-485e-4ec9-a08a-ff343118ce28 req-f9a5810a-07ff-4904-980c-0c85d858cd35 service nova] Acquiring lock "refresh_cache-43ee8948-8805-4d48-bd45-e93d2e2eb05d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.351412] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 850.351412] env[62914]: value = "task-1352561" [ 850.351412] env[62914]: _type = "Task" [ 850.351412] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.359401] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352561, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.373121] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 850.373121] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5259b650-5c52-7fc3-0bf3-6f49a39005ce" [ 850.373121] env[62914]: _type = "HttpNfcLease" [ 850.373121] env[62914]: } is ready. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 850.373476] env[62914]: DEBUG oslo_vmware.rw_handles [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 850.373476] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5259b650-5c52-7fc3-0bf3-6f49a39005ce" [ 850.373476] env[62914]: _type = "HttpNfcLease" [ 850.373476] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 850.374201] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c56732-cf17-4bd4-abb4-91df07c1e28c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.383207] env[62914]: DEBUG oslo_vmware.rw_handles [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f13b90-a948-8ccb-bd03-bb6e4920cc23/disk-0.vmdk from lease info. {{(pid=62914) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 850.383207] env[62914]: DEBUG oslo_vmware.rw_handles [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f13b90-a948-8ccb-bd03-bb6e4920cc23/disk-0.vmdk for reading. {{(pid=62914) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 850.554281] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-92f2aea0-5a18-45b2-862c-4509586689ab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.567820] env[62914]: DEBUG nova.network.neutron [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 850.730096] env[62914]: DEBUG oslo_concurrency.lockutils [req-8adc8037-623f-46e8-b6d0-f2050ea8413f req-fb11139b-b5e4-4994-a378-2ef23ad6713c service nova] Releasing lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.730463] env[62914]: DEBUG nova.compute.manager [req-8adc8037-623f-46e8-b6d0-f2050ea8413f req-fb11139b-b5e4-4994-a378-2ef23ad6713c service nova] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Received event network-vif-deleted-a5b49567-42ef-4e73-ab33-113aa689e83b {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.740515] env[62914]: DEBUG nova.network.neutron [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Updating instance_info_cache with network_info: [{"id": "cf125424-b194-49ac-886d-a8b053ae7a77", "address": "fa:16:3e:0f:7f:9c", "network": {"id": "c9c30d64-cafe-48a5-b868-b20d90d8920d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-458611021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5fa79f0260e461c952301f904101f79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf125424-b1", "ovs_interfaceid": "cf125424-b194-49ac-886d-a8b053ae7a77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.762133] env[62914]: DEBUG nova.scheduler.client.report [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 850.774656] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5208e8d9-2b67-44ed-04f3-106ae14e67b5, 'name': SearchDatastore_Task, 'duration_secs': 0.040826} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.776912] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.777191] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.777453] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.777599] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.777794] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.778267] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-debd9548-44ac-442f-b9c3-6d26a58c0e19 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.796947] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.797254] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 850.798607] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f480aa1-a83a-4322-af5c-8c4204add5a6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.805271] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 850.805271] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]528cfc87-5aee-fdd7-9cd2-10321dc9d5e2" [ 850.805271] env[62914]: _type = "Task" [ 850.805271] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.814992] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]528cfc87-5aee-fdd7-9cd2-10321dc9d5e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.862351] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352561, 'name': ReconfigVM_Task, 'duration_secs': 0.439428} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.862683] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Reconfigured VM instance instance-00000048 to attach disk [datastore1] c7cf71d7-b823-471f-949f-9ef2b539d972/c7cf71d7-b823-471f-949f-9ef2b539d972.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.863377] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-56fac60c-fb99-43b2-8a13-82d0d66ed44f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.870188] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 850.870188] env[62914]: value = "task-1352562" [ 850.870188] env[62914]: _type = "Task" [ 850.870188] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.878855] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352562, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.243331] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "refresh_cache-43ee8948-8805-4d48-bd45-e93d2e2eb05d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.243725] env[62914]: DEBUG nova.compute.manager [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Instance network_info: |[{"id": "cf125424-b194-49ac-886d-a8b053ae7a77", "address": "fa:16:3e:0f:7f:9c", "network": {"id": "c9c30d64-cafe-48a5-b868-b20d90d8920d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-458611021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5fa79f0260e461c952301f904101f79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf125424-b1", "ovs_interfaceid": "cf125424-b194-49ac-886d-a8b053ae7a77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 851.244096] env[62914]: DEBUG oslo_concurrency.lockutils [req-2a292e54-485e-4ec9-a08a-ff343118ce28 req-f9a5810a-07ff-4904-980c-0c85d858cd35 service nova] Acquired lock "refresh_cache-43ee8948-8805-4d48-bd45-e93d2e2eb05d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.244309] env[62914]: DEBUG nova.network.neutron [req-2a292e54-485e-4ec9-a08a-ff343118ce28 req-f9a5810a-07ff-4904-980c-0c85d858cd35 service nova] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Refreshing network info cache for port cf125424-b194-49ac-886d-a8b053ae7a77 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 851.245659] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0f:7f:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cf125424-b194-49ac-886d-a8b053ae7a77', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.253487] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Creating folder: Project (f5fa79f0260e461c952301f904101f79). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 851.254094] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a1738ca7-065f-4c40-bad3-fb4f2386f6f1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.267596] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Created folder: Project (f5fa79f0260e461c952301f904101f79) in parent group-v288131. [ 851.267825] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Creating folder: Instances. Parent ref: group-v288200. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 851.268190] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af15618a-5cb4-4b2c-9a46-83333f7470ce {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.270559] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.271144] env[62914]: DEBUG nova.compute.manager [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 851.273789] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.152s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.274177] env[62914]: DEBUG nova.objects.instance [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lazy-loading 'resources' on Instance uuid 9fec3e97-4203-4a90-8a43-600d11d7e7a0 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.284166] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Created folder: Instances in parent group-v288200. [ 851.284483] env[62914]: DEBUG oslo.service.loopingcall [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.285232] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 851.285232] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35f78112-cc32-4a52-8f55-aab3b949b152 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.312658] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 851.312658] env[62914]: value = "task-1352565" [ 851.312658] env[62914]: _type = "Task" [ 851.312658] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.320305] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]528cfc87-5aee-fdd7-9cd2-10321dc9d5e2, 'name': SearchDatastore_Task, 'duration_secs': 0.014274} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.321599] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69a5fd51-aab7-400d-9ce0-7235f63d76c0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.328295] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352565, 'name': CreateVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.331610] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 851.331610] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52af640b-f769-698c-642b-8f80788a537d" [ 851.331610] env[62914]: _type = "Task" [ 851.331610] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.340993] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52af640b-f769-698c-642b-8f80788a537d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.382834] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352562, 'name': Rename_Task, 'duration_secs': 0.159839} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.383360] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 851.384530] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd73992a-2888-47db-a692-86834c1f1a84 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.390184] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 851.390184] env[62914]: value = "task-1352566" [ 851.390184] env[62914]: _type = "Task" [ 851.390184] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.398690] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352566, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.778028] env[62914]: DEBUG nova.compute.utils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 851.783019] env[62914]: DEBUG nova.compute.manager [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 851.783299] env[62914]: DEBUG nova.network.neutron [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 851.823758] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352565, 'name': CreateVM_Task, 'duration_secs': 0.48942} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.826494] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.827611] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.827786] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.828127] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.828387] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01152c4f-ca87-4620-b7e0-9e47e9fb2dcb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.836814] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 851.836814] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]526b68af-1429-651d-bb97-4acf485623d6" [ 851.836814] env[62914]: _type = "Task" [ 851.836814] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.846412] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52af640b-f769-698c-642b-8f80788a537d, 'name': SearchDatastore_Task, 'duration_secs': 0.015362} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.847491] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.847731] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 4bf76ce0-b5d2-4184-888c-d0ef39878356/4bf76ce0-b5d2-4184-888c-d0ef39878356.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 851.850120] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-258c976a-dec9-4f19-888f-9a239dd57659 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.853164] env[62914]: DEBUG nova.policy [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a8a9795e30404c0d9969e658e767e015', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '28f67ce59c7f4151b37932bb890c4a5a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 851.858476] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]526b68af-1429-651d-bb97-4acf485623d6, 'name': SearchDatastore_Task, 'duration_secs': 0.016606} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.861585] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.861879] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.862147] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.862316] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.862507] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.863591] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3fb8fd7a-2e34-4dfd-9ec2-fb4d61a650c9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.866170] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 851.866170] env[62914]: value = "task-1352567" [ 851.866170] env[62914]: _type = "Task" [ 851.866170] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.875403] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352567, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.883336] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.883530] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 851.884410] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f93a6b87-ff05-40bc-9700-ad4095f4bde2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.890419] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 851.890419] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52180490-0c8f-3788-a5fd-577bc1e7642d" [ 851.890419] env[62914]: _type = "Task" [ 851.890419] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.910064] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52180490-0c8f-3788-a5fd-577bc1e7642d, 'name': SearchDatastore_Task, 'duration_secs': 0.011983} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.910411] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352566, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.914461] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-382e86dd-150a-4a3a-9c9a-18fd0996aac9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.918715] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 851.918715] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]526a5f7f-db8a-ebde-9141-67b5d0bfd7b9" [ 851.918715] env[62914]: _type = "Task" [ 851.918715] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.929095] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]526a5f7f-db8a-ebde-9141-67b5d0bfd7b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.033536] env[62914]: DEBUG nova.network.neutron [req-2a292e54-485e-4ec9-a08a-ff343118ce28 req-f9a5810a-07ff-4904-980c-0c85d858cd35 service nova] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Updated VIF entry in instance network info cache for port cf125424-b194-49ac-886d-a8b053ae7a77. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 852.033963] env[62914]: DEBUG nova.network.neutron [req-2a292e54-485e-4ec9-a08a-ff343118ce28 req-f9a5810a-07ff-4904-980c-0c85d858cd35 service nova] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Updating instance_info_cache with network_info: [{"id": "cf125424-b194-49ac-886d-a8b053ae7a77", "address": "fa:16:3e:0f:7f:9c", "network": {"id": "c9c30d64-cafe-48a5-b868-b20d90d8920d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-458611021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5fa79f0260e461c952301f904101f79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf125424-b1", "ovs_interfaceid": "cf125424-b194-49ac-886d-a8b053ae7a77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.043491] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20faef80-5d37-452c-9a86-4dba521bd38a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.053338] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20e846d-abb8-4641-b8b7-56e0ddd9b7ee {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.086179] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8843cb-741c-4be8-95a3-0e608a903285 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.094435] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb5c7c4-be58-4fd7-a56e-250af825a756 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.112082] env[62914]: DEBUG nova.compute.provider_tree [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.143777] env[62914]: DEBUG nova.network.neutron [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Successfully created port: bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.286485] env[62914]: DEBUG nova.compute.manager [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 852.378123] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352567, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.380842] env[62914]: DEBUG nova.compute.manager [req-57a4491b-2c24-47d4-99a5-548771e66e35 req-be6977f4-4582-4016-aa94-d64893d1cf68 service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Received event network-changed-59eaafaa-3160-4236-ab11-eb6f5e3ffec8 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.381173] env[62914]: DEBUG nova.compute.manager [req-57a4491b-2c24-47d4-99a5-548771e66e35 req-be6977f4-4582-4016-aa94-d64893d1cf68 service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Refreshing instance network info cache due to event network-changed-59eaafaa-3160-4236-ab11-eb6f5e3ffec8. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 852.381559] env[62914]: DEBUG oslo_concurrency.lockutils [req-57a4491b-2c24-47d4-99a5-548771e66e35 req-be6977f4-4582-4016-aa94-d64893d1cf68 service nova] Acquiring lock "refresh_cache-1fe15423-f8be-4763-b55e-2cbb383ff01d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.382195] env[62914]: DEBUG oslo_concurrency.lockutils [req-57a4491b-2c24-47d4-99a5-548771e66e35 req-be6977f4-4582-4016-aa94-d64893d1cf68 service nova] Acquired lock "refresh_cache-1fe15423-f8be-4763-b55e-2cbb383ff01d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.382195] env[62914]: DEBUG nova.network.neutron [req-57a4491b-2c24-47d4-99a5-548771e66e35 req-be6977f4-4582-4016-aa94-d64893d1cf68 service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Refreshing network info cache for port 59eaafaa-3160-4236-ab11-eb6f5e3ffec8 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 852.403392] env[62914]: DEBUG oslo_vmware.api [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352566, 'name': PowerOnVM_Task, 'duration_secs': 0.79169} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.403685] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 852.403883] env[62914]: INFO nova.compute.manager [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Took 9.14 seconds to spawn the instance on the hypervisor. [ 852.404106] env[62914]: DEBUG nova.compute.manager [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.404914] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a508b7ce-d959-48ee-a59f-5ed8de703df8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.430016] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]526a5f7f-db8a-ebde-9141-67b5d0bfd7b9, 'name': SearchDatastore_Task, 'duration_secs': 0.010796} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.430016] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.430016] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 43ee8948-8805-4d48-bd45-e93d2e2eb05d/43ee8948-8805-4d48-bd45-e93d2e2eb05d.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 852.430016] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d1746b9c-3280-4661-8b2a-997d51d23f2c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.435540] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 852.435540] env[62914]: value = "task-1352568" [ 852.435540] env[62914]: _type = "Task" [ 852.435540] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.443574] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352568, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.537192] env[62914]: DEBUG oslo_concurrency.lockutils [req-2a292e54-485e-4ec9-a08a-ff343118ce28 req-f9a5810a-07ff-4904-980c-0c85d858cd35 service nova] Releasing lock "refresh_cache-43ee8948-8805-4d48-bd45-e93d2e2eb05d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.620205] env[62914]: DEBUG nova.scheduler.client.report [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.878513] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352567, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575785} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.878858] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 4bf76ce0-b5d2-4184-888c-d0ef39878356/4bf76ce0-b5d2-4184-888c-d0ef39878356.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 852.879045] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.879484] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-94812a22-c363-4462-85c8-3477807a67d9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.887655] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 852.887655] env[62914]: value = "task-1352569" [ 852.887655] env[62914]: _type = "Task" [ 852.887655] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.895983] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352569, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.926962] env[62914]: INFO nova.compute.manager [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Took 23.48 seconds to build instance. [ 852.947756] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352568, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.125747] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.852s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.128864] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.672s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.131650] env[62914]: INFO nova.compute.claims [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 853.135989] env[62914]: DEBUG nova.network.neutron [req-57a4491b-2c24-47d4-99a5-548771e66e35 req-be6977f4-4582-4016-aa94-d64893d1cf68 service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Updated VIF entry in instance network info cache for port 59eaafaa-3160-4236-ab11-eb6f5e3ffec8. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 853.136460] env[62914]: DEBUG nova.network.neutron [req-57a4491b-2c24-47d4-99a5-548771e66e35 req-be6977f4-4582-4016-aa94-d64893d1cf68 service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Updating instance_info_cache with network_info: [{"id": "59eaafaa-3160-4236-ab11-eb6f5e3ffec8", "address": "fa:16:3e:aa:2e:4c", "network": {"id": "184ac684-c4e6-4468-aee8-1625d96dc0b4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1513916205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3055903b6f4443bbeb9897e75e34c49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59eaafaa-31", "ovs_interfaceid": "59eaafaa-3160-4236-ab11-eb6f5e3ffec8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.153402] env[62914]: INFO nova.scheduler.client.report [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleted allocations for instance 9fec3e97-4203-4a90-8a43-600d11d7e7a0 [ 853.295792] env[62914]: DEBUG nova.compute.manager [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 853.317171] env[62914]: DEBUG nova.virt.hardware [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 853.317488] env[62914]: DEBUG nova.virt.hardware [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 853.317666] env[62914]: DEBUG nova.virt.hardware [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.317853] env[62914]: DEBUG nova.virt.hardware [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 853.317998] env[62914]: DEBUG nova.virt.hardware [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.318158] env[62914]: DEBUG nova.virt.hardware [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 853.318364] env[62914]: DEBUG nova.virt.hardware [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 853.318547] env[62914]: DEBUG nova.virt.hardware [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 853.318730] env[62914]: DEBUG nova.virt.hardware [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 853.318897] env[62914]: DEBUG nova.virt.hardware [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 853.319088] env[62914]: DEBUG nova.virt.hardware [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 853.320034] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bcde218-7e4a-45a4-bdf1-5ca18822296d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.327743] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0336ff7-102d-4153-b6a8-5e75c98bef95 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.397192] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352569, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.263338} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.397571] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 853.398425] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3dbcd5a-de68-4dda-bab2-6e14403461b3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.421375] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 4bf76ce0-b5d2-4184-888c-d0ef39878356/4bf76ce0-b5d2-4184-888c-d0ef39878356.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 853.421670] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-759e3840-7f49-4e54-9bbc-96280cb7965c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.436306] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c52e31e5-145f-4de1-9b7d-5cf0325ac6f9 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.634s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.445149] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352568, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.622301} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.446398] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 43ee8948-8805-4d48-bd45-e93d2e2eb05d/43ee8948-8805-4d48-bd45-e93d2e2eb05d.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 853.446616] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 853.446991] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 853.446991] env[62914]: value = "task-1352570" [ 853.446991] env[62914]: _type = "Task" [ 853.446991] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.447235] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b3badb52-02a5-4fec-a51d-fcff7becd6e3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.458247] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352570, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.459529] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 853.459529] env[62914]: value = "task-1352571" [ 853.459529] env[62914]: _type = "Task" [ 853.459529] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.467029] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352571, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.642126] env[62914]: DEBUG oslo_concurrency.lockutils [req-57a4491b-2c24-47d4-99a5-548771e66e35 req-be6977f4-4582-4016-aa94-d64893d1cf68 service nova] Releasing lock "refresh_cache-1fe15423-f8be-4763-b55e-2cbb383ff01d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.661801] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a60b5bb1-9598-4725-9723-6951c41ef6e8 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "9fec3e97-4203-4a90-8a43-600d11d7e7a0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.424s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.711356] env[62914]: DEBUG nova.compute.manager [req-7aeb2b0a-65ff-44b6-ba96-ac48a7705a09 req-334c04fc-f99c-48a2-bb9f-4193cc560eb8 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Received event network-vif-plugged-bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.714236] env[62914]: DEBUG oslo_concurrency.lockutils [req-7aeb2b0a-65ff-44b6-ba96-ac48a7705a09 req-334c04fc-f99c-48a2-bb9f-4193cc560eb8 service nova] Acquiring lock "29680cfb-8df3-4ff7-b6ee-0982577cc9cf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.714236] env[62914]: DEBUG oslo_concurrency.lockutils [req-7aeb2b0a-65ff-44b6-ba96-ac48a7705a09 req-334c04fc-f99c-48a2-bb9f-4193cc560eb8 service nova] Lock "29680cfb-8df3-4ff7-b6ee-0982577cc9cf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.714236] env[62914]: DEBUG oslo_concurrency.lockutils [req-7aeb2b0a-65ff-44b6-ba96-ac48a7705a09 req-334c04fc-f99c-48a2-bb9f-4193cc560eb8 service nova] Lock "29680cfb-8df3-4ff7-b6ee-0982577cc9cf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.714236] env[62914]: DEBUG nova.compute.manager [req-7aeb2b0a-65ff-44b6-ba96-ac48a7705a09 req-334c04fc-f99c-48a2-bb9f-4193cc560eb8 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] No waiting events found dispatching network-vif-plugged-bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 853.714236] env[62914]: WARNING nova.compute.manager [req-7aeb2b0a-65ff-44b6-ba96-ac48a7705a09 req-334c04fc-f99c-48a2-bb9f-4193cc560eb8 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Received unexpected event network-vif-plugged-bf4a3fa8-08a9-4890-abf6-fa02a60168a2 for instance with vm_state building and task_state spawning. [ 853.798591] env[62914]: DEBUG nova.network.neutron [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Successfully updated port: bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 853.963516] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.976153] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352571, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.195262} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.976461] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 853.977465] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b09adf-ae48-4055-9708-08747c092de5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.000071] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 43ee8948-8805-4d48-bd45-e93d2e2eb05d/43ee8948-8805-4d48-bd45-e93d2e2eb05d.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.000390] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bbedf786-243d-478b-a908-e1d24c87ee95 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.027288] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 854.027288] env[62914]: value = "task-1352572" [ 854.027288] env[62914]: _type = "Task" [ 854.027288] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.031936] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352572, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.302420] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquiring lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.302420] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquired lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.302420] env[62914]: DEBUG nova.network.neutron [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 854.355085] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a7c06f-0dee-4034-9050-7aa6858d2937 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.363030] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49473464-dadb-4d8f-8d88-4fc97b1f09c6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.397102] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b5aaea-6ad2-4748-8442-7ccc1d9537a3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.404999] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c67e590-4623-42ab-a458-90409eb7932e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.411824] env[62914]: DEBUG nova.compute.manager [req-dc20f46a-291d-4b30-8d7a-aae5e7102f0c req-410df459-0b98-4ab0-a47d-689ff8a37c81 service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Received event network-changed-57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.412061] env[62914]: DEBUG nova.compute.manager [req-dc20f46a-291d-4b30-8d7a-aae5e7102f0c req-410df459-0b98-4ab0-a47d-689ff8a37c81 service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Refreshing instance network info cache due to event network-changed-57172af7-5ed6-46d9-9d7b-3ef100c530f7. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 854.412292] env[62914]: DEBUG oslo_concurrency.lockutils [req-dc20f46a-291d-4b30-8d7a-aae5e7102f0c req-410df459-0b98-4ab0-a47d-689ff8a37c81 service nova] Acquiring lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.412452] env[62914]: DEBUG oslo_concurrency.lockutils [req-dc20f46a-291d-4b30-8d7a-aae5e7102f0c req-410df459-0b98-4ab0-a47d-689ff8a37c81 service nova] Acquired lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.412612] env[62914]: DEBUG nova.network.neutron [req-dc20f46a-291d-4b30-8d7a-aae5e7102f0c req-410df459-0b98-4ab0-a47d-689ff8a37c81 service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Refreshing network info cache for port 57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 854.428791] env[62914]: DEBUG nova.compute.provider_tree [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.461236] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352570, 'name': ReconfigVM_Task, 'duration_secs': 0.684405} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.461546] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 4bf76ce0-b5d2-4184-888c-d0ef39878356/4bf76ce0-b5d2-4184-888c-d0ef39878356.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.462327] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-40a036fe-ba35-4fbe-a11e-309e4945390e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.468445] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 854.468445] env[62914]: value = "task-1352573" [ 854.468445] env[62914]: _type = "Task" [ 854.468445] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.476659] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352573, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.534729] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352572, 'name': ReconfigVM_Task, 'duration_secs': 0.472463} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.534978] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 43ee8948-8805-4d48-bd45-e93d2e2eb05d/43ee8948-8805-4d48-bd45-e93d2e2eb05d.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.535733] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b4b448be-772d-403b-b0d0-53d2cf0fb535 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.542942] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 854.542942] env[62914]: value = "task-1352574" [ 854.542942] env[62914]: _type = "Task" [ 854.542942] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.551944] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352574, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.914608] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "22de53a6-096b-4e8e-af91-f78588ea4c60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.914862] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "22de53a6-096b-4e8e-af91-f78588ea4c60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.934112] env[62914]: DEBUG nova.scheduler.client.report [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.981165] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352573, 'name': Rename_Task, 'duration_secs': 0.243384} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.981540] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 854.981811] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f4dfd020-b0f4-4ec4-9555-96281f286e5f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.988871] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 854.988871] env[62914]: value = "task-1352575" [ 854.988871] env[62914]: _type = "Task" [ 854.988871] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.997020] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352575, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.042789] env[62914]: DEBUG nova.network.neutron [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 855.057610] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352574, 'name': Rename_Task, 'duration_secs': 0.137573} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.058040] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.060917] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d3a9485-5e0c-4768-a9e8-96df9af33a0d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.067501] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 855.067501] env[62914]: value = "task-1352576" [ 855.067501] env[62914]: _type = "Task" [ 855.067501] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.079814] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352576, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.194126] env[62914]: DEBUG nova.network.neutron [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Updating instance_info_cache with network_info: [{"id": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "address": "fa:16:3e:c9:b5:2c", "network": {"id": "16e49b25-5bce-4cf2-8199-a865e6cd483e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1094957463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "28f67ce59c7f4151b37932bb890c4a5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf4a3fa8-08", "ovs_interfaceid": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.265565] env[62914]: DEBUG nova.network.neutron [req-dc20f46a-291d-4b30-8d7a-aae5e7102f0c req-410df459-0b98-4ab0-a47d-689ff8a37c81 service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updated VIF entry in instance network info cache for port 57172af7-5ed6-46d9-9d7b-3ef100c530f7. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 855.266214] env[62914]: DEBUG nova.network.neutron [req-dc20f46a-291d-4b30-8d7a-aae5e7102f0c req-410df459-0b98-4ab0-a47d-689ff8a37c81 service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updating instance_info_cache with network_info: [{"id": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "address": "fa:16:3e:91:13:35", "network": {"id": "c4298e11-ab16-4b37-be29-48fc406e74b0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-695272838-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.236", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "15315f8e77d749a4b2b0211c3b22b546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe38bb7e-8bcb-419d-868f-0dc105c69651", "external-id": "nsx-vlan-transportzone-432", "segmentation_id": 432, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57172af7-5e", "ovs_interfaceid": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.422443] env[62914]: DEBUG nova.compute.manager [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 855.441038] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.441761] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.608s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.442039] env[62914]: DEBUG nova.objects.instance [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lazy-loading 'resources' on Instance uuid 46d139a1-a4ae-435e-9d6f-cfc06d706128 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 855.499567] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352575, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.581067] env[62914]: DEBUG oslo_vmware.api [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352576, 'name': PowerOnVM_Task, 'duration_secs': 0.50708} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.581067] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 855.581223] env[62914]: INFO nova.compute.manager [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Took 6.46 seconds to spawn the instance on the hypervisor. [ 855.581329] env[62914]: DEBUG nova.compute.manager [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.582271] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d75d36b-611b-4598-ac40-d242b4ec1d3c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.696469] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Releasing lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.696782] env[62914]: DEBUG nova.compute.manager [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Instance network_info: |[{"id": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "address": "fa:16:3e:c9:b5:2c", "network": {"id": "16e49b25-5bce-4cf2-8199-a865e6cd483e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1094957463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "28f67ce59c7f4151b37932bb890c4a5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf4a3fa8-08", "ovs_interfaceid": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 855.697299] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:b5:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9f208df-1fb5-4403-9796-7fd19e4bfb85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bf4a3fa8-08a9-4890-abf6-fa02a60168a2', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.705103] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Creating folder: Project (28f67ce59c7f4151b37932bb890c4a5a). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.705791] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3b7cf7bc-ef13-4ba2-8c32-57e3125da02b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.720987] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Created folder: Project (28f67ce59c7f4151b37932bb890c4a5a) in parent group-v288131. [ 855.721204] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Creating folder: Instances. Parent ref: group-v288203. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.721461] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1be955b0-a801-4997-a061-763d7e73ff5c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.732358] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Created folder: Instances in parent group-v288203. [ 855.732648] env[62914]: DEBUG oslo.service.loopingcall [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.732862] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 855.733076] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-84ac5d86-f6e5-42eb-b64c-ad0e8cef1631 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.749819] env[62914]: DEBUG nova.compute.manager [req-246142e6-e549-4ad9-8aa9-fe7d9dc59488 req-20e8019f-0582-4ec3-abcd-cbcb7ae21764 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Received event network-changed-bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.750021] env[62914]: DEBUG nova.compute.manager [req-246142e6-e549-4ad9-8aa9-fe7d9dc59488 req-20e8019f-0582-4ec3-abcd-cbcb7ae21764 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Refreshing instance network info cache due to event network-changed-bf4a3fa8-08a9-4890-abf6-fa02a60168a2. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 855.750248] env[62914]: DEBUG oslo_concurrency.lockutils [req-246142e6-e549-4ad9-8aa9-fe7d9dc59488 req-20e8019f-0582-4ec3-abcd-cbcb7ae21764 service nova] Acquiring lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.750391] env[62914]: DEBUG oslo_concurrency.lockutils [req-246142e6-e549-4ad9-8aa9-fe7d9dc59488 req-20e8019f-0582-4ec3-abcd-cbcb7ae21764 service nova] Acquired lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.750629] env[62914]: DEBUG nova.network.neutron [req-246142e6-e549-4ad9-8aa9-fe7d9dc59488 req-20e8019f-0582-4ec3-abcd-cbcb7ae21764 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Refreshing network info cache for port bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 855.756301] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.756301] env[62914]: value = "task-1352579" [ 855.756301] env[62914]: _type = "Task" [ 855.756301] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.763856] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352579, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.769650] env[62914]: DEBUG oslo_concurrency.lockutils [req-dc20f46a-291d-4b30-8d7a-aae5e7102f0c req-410df459-0b98-4ab0-a47d-689ff8a37c81 service nova] Releasing lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.945094] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Acquiring lock "f9f42506-4133-4dff-9a8d-71508d6280b7" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.945511] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Lock "f9f42506-4133-4dff-9a8d-71508d6280b7" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.952024] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.999442] env[62914]: DEBUG oslo_vmware.api [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352575, 'name': PowerOnVM_Task, 'duration_secs': 0.765412} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.002358] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.002581] env[62914]: INFO nova.compute.manager [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Took 9.24 seconds to spawn the instance on the hypervisor. [ 856.002772] env[62914]: DEBUG nova.compute.manager [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.003929] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90461ba7-01b4-4fd4-9a48-3c7d33cff558 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.102749] env[62914]: INFO nova.compute.manager [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Took 21.97 seconds to build instance. [ 856.170024] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69379623-f693-4a28-bd64-4f3d74f60394 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.178618] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ee1fc9-aa3f-42d6-ba33-c6498ac361d8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.215761] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937341d3-dfb5-4fd9-9b2c-528650d6ad5f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.224239] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1721e95-509f-45ff-8e7f-540d8f8b31f7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.238382] env[62914]: DEBUG nova.compute.provider_tree [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.265073] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352579, 'name': CreateVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.451297] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Lock "f9f42506-4133-4dff-9a8d-71508d6280b7" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.506s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.451907] env[62914]: DEBUG nova.compute.manager [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 856.475831] env[62914]: DEBUG nova.network.neutron [req-246142e6-e549-4ad9-8aa9-fe7d9dc59488 req-20e8019f-0582-4ec3-abcd-cbcb7ae21764 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Updated VIF entry in instance network info cache for port bf4a3fa8-08a9-4890-abf6-fa02a60168a2. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 856.476358] env[62914]: DEBUG nova.network.neutron [req-246142e6-e549-4ad9-8aa9-fe7d9dc59488 req-20e8019f-0582-4ec3-abcd-cbcb7ae21764 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Updating instance_info_cache with network_info: [{"id": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "address": "fa:16:3e:c9:b5:2c", "network": {"id": "16e49b25-5bce-4cf2-8199-a865e6cd483e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1094957463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "28f67ce59c7f4151b37932bb890c4a5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf4a3fa8-08", "ovs_interfaceid": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.521656] env[62914]: INFO nova.compute.manager [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Took 25.20 seconds to build instance. [ 856.607296] env[62914]: DEBUG oslo_concurrency.lockutils [None req-89c19dde-1903-4802-8975-52ce0a8a0063 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.413s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.744025] env[62914]: DEBUG nova.scheduler.client.report [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.768215] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352579, 'name': CreateVM_Task, 'duration_secs': 0.558358} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.768426] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 856.769171] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.769385] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.770038] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.770125] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07a5bae4-b223-4ca2-add9-f1c07352cb00 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.777906] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 856.777906] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52058bf0-0c8a-b98a-b6e3-ba3945313447" [ 856.777906] env[62914]: _type = "Task" [ 856.777906] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.788235] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52058bf0-0c8a-b98a-b6e3-ba3945313447, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.957508] env[62914]: DEBUG nova.compute.utils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 856.958989] env[62914]: DEBUG nova.compute.manager [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 856.959370] env[62914]: DEBUG nova.network.neutron [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 856.979577] env[62914]: DEBUG oslo_concurrency.lockutils [req-246142e6-e549-4ad9-8aa9-fe7d9dc59488 req-20e8019f-0582-4ec3-abcd-cbcb7ae21764 service nova] Releasing lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.024602] env[62914]: DEBUG oslo_concurrency.lockutils [None req-373c94b6-c5d9-469f-a756-38c7f507db8a tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "4bf76ce0-b5d2-4184-888c-d0ef39878356" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.937s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.027110] env[62914]: DEBUG nova.policy [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '892540cf7f5a4703a36800627cb0e475', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b4250bae94c426da87ef1a8d45524d5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 857.248020] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.806s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.251032] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.112s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.251287] env[62914]: DEBUG nova.objects.instance [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lazy-loading 'resources' on Instance uuid cb89c84f-414b-4dc7-9db5-12cbc30b52b4 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.270704] env[62914]: INFO nova.scheduler.client.report [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Deleted allocations for instance 46d139a1-a4ae-435e-9d6f-cfc06d706128 [ 857.288305] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52058bf0-0c8a-b98a-b6e3-ba3945313447, 'name': SearchDatastore_Task, 'duration_secs': 0.016545} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.288533] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.288766] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.289063] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.289245] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.289433] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.289742] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-876548c7-7458-42d9-9e2c-bc9c523db752 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.303207] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.303207] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 857.303599] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f16a2b17-55fe-4ebe-ade3-18b8b796bfa2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.310173] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 857.310173] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52b4d143-69e3-c16b-8e14-53f0661657f3" [ 857.310173] env[62914]: _type = "Task" [ 857.310173] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.320045] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52b4d143-69e3-c16b-8e14-53f0661657f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.320827] env[62914]: DEBUG nova.network.neutron [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Successfully created port: f1c9a0bc-9d38-477f-a38e-75237f989567 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 857.419012] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b1fdbf44-0d5d-4bf8-926c-ffcba6b7b4e7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.419316] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b1fdbf44-0d5d-4bf8-926c-ffcba6b7b4e7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.419700] env[62914]: DEBUG nova.compute.manager [None req-b1fdbf44-0d5d-4bf8-926c-ffcba6b7b4e7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.420967] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90499569-eb65-49bb-9870-4c21c0094aff {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.428559] env[62914]: DEBUG nova.compute.manager [None req-b1fdbf44-0d5d-4bf8-926c-ffcba6b7b4e7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62914) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 857.429203] env[62914]: DEBUG nova.objects.instance [None req-b1fdbf44-0d5d-4bf8-926c-ffcba6b7b4e7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lazy-loading 'flavor' on Instance uuid 43ee8948-8805-4d48-bd45-e93d2e2eb05d {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.462590] env[62914]: DEBUG nova.compute.manager [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 857.786645] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4a97fbac-1c72-4896-ab1e-7cfa8a81cc81 tempest-SecurityGroupsTestJSON-422525159 tempest-SecurityGroupsTestJSON-422525159-project-member] Lock "46d139a1-a4ae-435e-9d6f-cfc06d706128" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.959s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.826513] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52b4d143-69e3-c16b-8e14-53f0661657f3, 'name': SearchDatastore_Task, 'duration_secs': 0.017736} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.827285] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2b8a6f6-da6b-4c2a-8e5e-94765adab0c9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.838779] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 857.838779] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5286eb56-f4a6-5aea-83e6-55237b7bf5ae" [ 857.838779] env[62914]: _type = "Task" [ 857.838779] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.857893] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5286eb56-f4a6-5aea-83e6-55237b7bf5ae, 'name': SearchDatastore_Task, 'duration_secs': 0.015846} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.857893] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.857893] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 29680cfb-8df3-4ff7-b6ee-0982577cc9cf/29680cfb-8df3-4ff7-b6ee-0982577cc9cf.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 857.858293] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b0a18132-e294-46bc-9416-a6daf00b3971 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.865777] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 857.865777] env[62914]: value = "task-1352580" [ 857.865777] env[62914]: _type = "Task" [ 857.865777] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.874590] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352580, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.897955] env[62914]: DEBUG nova.compute.manager [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Stashing vm_state: active {{(pid=62914) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 857.935318] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1fdbf44-0d5d-4bf8-926c-ffcba6b7b4e7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 857.935585] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8652b4f7-3ab4-4bdd-81e0-6b3606e2dfd0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.943348] env[62914]: DEBUG oslo_vmware.api [None req-b1fdbf44-0d5d-4bf8-926c-ffcba6b7b4e7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 857.943348] env[62914]: value = "task-1352581" [ 857.943348] env[62914]: _type = "Task" [ 857.943348] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.953157] env[62914]: DEBUG oslo_vmware.api [None req-b1fdbf44-0d5d-4bf8-926c-ffcba6b7b4e7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352581, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.021191] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8abe0cc-6263-4a14-aa29-b16fa4a95e4c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.029997] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36fb9194-e975-43ed-a632-dccef58c504f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.062544] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba06efa8-460e-4b36-b566-cac4d7c36e70 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.070767] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5fc702-ea2a-43ad-84ba-94594d1d4c87 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.085061] env[62914]: DEBUG nova.compute.provider_tree [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.378776] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352580, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.416720] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.456420] env[62914]: DEBUG oslo_vmware.api [None req-b1fdbf44-0d5d-4bf8-926c-ffcba6b7b4e7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352581, 'name': PowerOffVM_Task, 'duration_secs': 0.194731} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.456867] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1fdbf44-0d5d-4bf8-926c-ffcba6b7b4e7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 858.456986] env[62914]: DEBUG nova.compute.manager [None req-b1fdbf44-0d5d-4bf8-926c-ffcba6b7b4e7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.457861] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a4efdf-4bb0-457d-be02-2c21ecc6b1d0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.471506] env[62914]: DEBUG nova.compute.manager [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 858.495653] env[62914]: DEBUG nova.virt.hardware [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.495941] env[62914]: DEBUG nova.virt.hardware [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.496295] env[62914]: DEBUG nova.virt.hardware [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.496367] env[62914]: DEBUG nova.virt.hardware [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.496463] env[62914]: DEBUG nova.virt.hardware [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.496609] env[62914]: DEBUG nova.virt.hardware [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.496846] env[62914]: DEBUG nova.virt.hardware [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.497037] env[62914]: DEBUG nova.virt.hardware [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.497290] env[62914]: DEBUG nova.virt.hardware [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.497485] env[62914]: DEBUG nova.virt.hardware [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.497999] env[62914]: DEBUG nova.virt.hardware [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.498911] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e839db-78b4-40e5-bccb-a2bdc32cce1b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.508793] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e248ad81-6411-46a2-ab14-47aa31d1aabd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.593089] env[62914]: DEBUG nova.scheduler.client.report [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.879076] env[62914]: DEBUG nova.compute.manager [req-8f169502-4a15-4de2-9c84-d4dcea77e29c req-497220d3-0a29-4aed-b132-9aaccf5b6167 service nova] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Received event network-vif-plugged-f1c9a0bc-9d38-477f-a38e-75237f989567 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.879252] env[62914]: DEBUG oslo_concurrency.lockutils [req-8f169502-4a15-4de2-9c84-d4dcea77e29c req-497220d3-0a29-4aed-b132-9aaccf5b6167 service nova] Acquiring lock "f80ee46d-9209-4b60-9e5a-f7c8764f4ccb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.879485] env[62914]: DEBUG oslo_concurrency.lockutils [req-8f169502-4a15-4de2-9c84-d4dcea77e29c req-497220d3-0a29-4aed-b132-9aaccf5b6167 service nova] Lock "f80ee46d-9209-4b60-9e5a-f7c8764f4ccb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.879699] env[62914]: DEBUG oslo_concurrency.lockutils [req-8f169502-4a15-4de2-9c84-d4dcea77e29c req-497220d3-0a29-4aed-b132-9aaccf5b6167 service nova] Lock "f80ee46d-9209-4b60-9e5a-f7c8764f4ccb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.879879] env[62914]: DEBUG nova.compute.manager [req-8f169502-4a15-4de2-9c84-d4dcea77e29c req-497220d3-0a29-4aed-b132-9aaccf5b6167 service nova] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] No waiting events found dispatching network-vif-plugged-f1c9a0bc-9d38-477f-a38e-75237f989567 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 858.880060] env[62914]: WARNING nova.compute.manager [req-8f169502-4a15-4de2-9c84-d4dcea77e29c req-497220d3-0a29-4aed-b132-9aaccf5b6167 service nova] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Received unexpected event network-vif-plugged-f1c9a0bc-9d38-477f-a38e-75237f989567 for instance with vm_state building and task_state spawning. [ 858.884191] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352580, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.672299} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.884438] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 29680cfb-8df3-4ff7-b6ee-0982577cc9cf/29680cfb-8df3-4ff7-b6ee-0982577cc9cf.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 858.884675] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 858.884934] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-96d5abb2-7a30-440c-b0cf-4dbc061dfa8c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.892953] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 858.892953] env[62914]: value = "task-1352582" [ 858.892953] env[62914]: _type = "Task" [ 858.892953] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.906127] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352582, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.939883] env[62914]: DEBUG nova.network.neutron [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Successfully updated port: f1c9a0bc-9d38-477f-a38e-75237f989567 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 858.970581] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b1fdbf44-0d5d-4bf8-926c-ffcba6b7b4e7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.551s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.098924] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.848s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.102339] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.821s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.102587] env[62914]: DEBUG nova.objects.instance [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lazy-loading 'resources' on Instance uuid 9577394d-1c73-4ed1-ba86-e7c246e32719 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.122765] env[62914]: INFO nova.scheduler.client.report [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Deleted allocations for instance cb89c84f-414b-4dc7-9db5-12cbc30b52b4 [ 859.410915] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352582, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076117} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.411260] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.412079] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f499876-1767-4ac3-a6cc-ca875381e391 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.438016] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 29680cfb-8df3-4ff7-b6ee-0982577cc9cf/29680cfb-8df3-4ff7-b6ee-0982577cc9cf.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.438894] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3de71d78-d4fc-43db-a539-228338e0384f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.457415] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Acquiring lock "refresh_cache-f80ee46d-9209-4b60-9e5a-f7c8764f4ccb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.457739] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Acquired lock "refresh_cache-f80ee46d-9209-4b60-9e5a-f7c8764f4ccb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.457897] env[62914]: DEBUG nova.network.neutron [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 859.464035] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 859.464035] env[62914]: value = "task-1352583" [ 859.464035] env[62914]: _type = "Task" [ 859.464035] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.474673] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352583, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.637293] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc149611-a115-44df-9f89-51b25fda03d9 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "cb89c84f-414b-4dc7-9db5-12cbc30b52b4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.456s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.839841] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f572688c-0880-42a1-836e-6c60e4da1b3d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.851729] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-927755b4-f1ac-492c-a920-186eb150da85 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.882508] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a2dad8-2225-458c-b983-2e32ee2f6493 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.891467] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002cbd1a-327a-4642-b0f8-12f162efc9f6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.910661] env[62914]: DEBUG nova.compute.provider_tree [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.974858] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352583, 'name': ReconfigVM_Task, 'duration_secs': 0.481105} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.975360] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 29680cfb-8df3-4ff7-b6ee-0982577cc9cf/29680cfb-8df3-4ff7-b6ee-0982577cc9cf.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.975875] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-25cdbbc3-71e4-424d-bc74-07641a5e87f1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.983637] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 859.983637] env[62914]: value = "task-1352584" [ 859.983637] env[62914]: _type = "Task" [ 859.983637] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.995263] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352584, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.997109] env[62914]: DEBUG nova.network.neutron [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.195273] env[62914]: DEBUG nova.network.neutron [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Updating instance_info_cache with network_info: [{"id": "f1c9a0bc-9d38-477f-a38e-75237f989567", "address": "fa:16:3e:2b:fa:04", "network": {"id": "85e33a61-2e23-4b42-822f-b277a00fd239", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-719568965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b4250bae94c426da87ef1a8d45524d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1c9a0bc-9d", "ovs_interfaceid": "f1c9a0bc-9d38-477f-a38e-75237f989567", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.415453] env[62914]: DEBUG nova.scheduler.client.report [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.495113] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352584, 'name': Rename_Task, 'duration_secs': 0.262815} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.495448] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 860.496064] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0aa8ed62-4dc1-484c-87d6-0b8d6a5828a5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.503152] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 860.503152] env[62914]: value = "task-1352585" [ 860.503152] env[62914]: _type = "Task" [ 860.503152] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.510291] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352585, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.697665] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Releasing lock "refresh_cache-f80ee46d-9209-4b60-9e5a-f7c8764f4ccb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.697998] env[62914]: DEBUG nova.compute.manager [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Instance network_info: |[{"id": "f1c9a0bc-9d38-477f-a38e-75237f989567", "address": "fa:16:3e:2b:fa:04", "network": {"id": "85e33a61-2e23-4b42-822f-b277a00fd239", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-719568965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b4250bae94c426da87ef1a8d45524d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1c9a0bc-9d", "ovs_interfaceid": "f1c9a0bc-9d38-477f-a38e-75237f989567", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 860.698455] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:fa:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '21310d90-efbc-45a8-a97f-c4358606530f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f1c9a0bc-9d38-477f-a38e-75237f989567', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 860.708132] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Creating folder: Project (0b4250bae94c426da87ef1a8d45524d5). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 860.708559] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a2dc6942-e1df-4557-ab5b-bde724b57c2e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.719904] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Created folder: Project (0b4250bae94c426da87ef1a8d45524d5) in parent group-v288131. [ 860.720212] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Creating folder: Instances. Parent ref: group-v288206. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 860.720547] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f264e2cd-72b5-479d-8b85-7e9b9a78fd02 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.730394] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Created folder: Instances in parent group-v288206. [ 860.730707] env[62914]: DEBUG oslo.service.loopingcall [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.730986] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 860.731283] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-54d20e13-4fd0-48c2-b876-24b514bcae86 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.759987] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 860.759987] env[62914]: value = "task-1352588" [ 860.759987] env[62914]: _type = "Task" [ 860.759987] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.768054] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352588, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.841767] env[62914]: DEBUG nova.compute.manager [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 860.842614] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d217e65d-87f0-40a7-bc56-2c801a164adf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.910875] env[62914]: DEBUG nova.compute.manager [req-5c0b135d-d783-46c6-95cc-b22fca877835 req-a47608bc-3d53-4809-b0a8-2da0b6e707ad service nova] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Received event network-changed-f1c9a0bc-9d38-477f-a38e-75237f989567 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.911246] env[62914]: DEBUG nova.compute.manager [req-5c0b135d-d783-46c6-95cc-b22fca877835 req-a47608bc-3d53-4809-b0a8-2da0b6e707ad service nova] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Refreshing instance network info cache due to event network-changed-f1c9a0bc-9d38-477f-a38e-75237f989567. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 860.911505] env[62914]: DEBUG oslo_concurrency.lockutils [req-5c0b135d-d783-46c6-95cc-b22fca877835 req-a47608bc-3d53-4809-b0a8-2da0b6e707ad service nova] Acquiring lock "refresh_cache-f80ee46d-9209-4b60-9e5a-f7c8764f4ccb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.911716] env[62914]: DEBUG oslo_concurrency.lockutils [req-5c0b135d-d783-46c6-95cc-b22fca877835 req-a47608bc-3d53-4809-b0a8-2da0b6e707ad service nova] Acquired lock "refresh_cache-f80ee46d-9209-4b60-9e5a-f7c8764f4ccb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.911839] env[62914]: DEBUG nova.network.neutron [req-5c0b135d-d783-46c6-95cc-b22fca877835 req-a47608bc-3d53-4809-b0a8-2da0b6e707ad service nova] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Refreshing network info cache for port f1c9a0bc-9d38-477f-a38e-75237f989567 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 860.919728] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.817s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.922588] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.357s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.922820] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.922932] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62914) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 860.923258] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.811s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.923477] env[62914]: DEBUG nova.objects.instance [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Lazy-loading 'resources' on Instance uuid bd3e82ce-62e7-4761-bcd7-d1f835234b97 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.926753] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9fef3d1-1aba-421f-b259-958e56719edc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.938131] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadb2044-7ebd-4f10-899b-275b8faf5d9d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.954341] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d52dbf0-709f-4371-bcd8-95bd5fe8fe47 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.958392] env[62914]: INFO nova.scheduler.client.report [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Deleted allocations for instance 9577394d-1c73-4ed1-ba86-e7c246e32719 [ 860.968255] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b35a29-0e12-4f0f-97e0-ffbe16341caa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.007537] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180885MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62914) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 861.007719] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.018953] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352585, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.040410] env[62914]: DEBUG oslo_vmware.rw_handles [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f13b90-a948-8ccb-bd03-bb6e4920cc23/disk-0.vmdk. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 861.041357] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28074e9f-5adc-46ae-a0cc-90de531d3692 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.048741] env[62914]: DEBUG oslo_vmware.rw_handles [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f13b90-a948-8ccb-bd03-bb6e4920cc23/disk-0.vmdk is in state: ready. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 861.049119] env[62914]: ERROR oslo_vmware.rw_handles [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f13b90-a948-8ccb-bd03-bb6e4920cc23/disk-0.vmdk due to incomplete transfer. [ 861.049479] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a78daa64-e5d3-48cf-ab5d-a7e371e4969d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.057943] env[62914]: DEBUG oslo_vmware.rw_handles [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f13b90-a948-8ccb-bd03-bb6e4920cc23/disk-0.vmdk. {{(pid=62914) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 861.058307] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Uploaded image 1040c236-a1c7-4558-88d1-f521d06e9843 to the Glance image server {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 861.061634] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Destroying the VM {{(pid=62914) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 861.062069] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e025b53a-89fb-419f-8bf9-83a0bb123dcc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.069276] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 861.069276] env[62914]: value = "task-1352589" [ 861.069276] env[62914]: _type = "Task" [ 861.069276] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.081452] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352589, 'name': Destroy_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.270618] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352588, 'name': CreateVM_Task, 'duration_secs': 0.467318} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.270896] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 861.271668] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.271827] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.272186] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 861.272446] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5da3b41f-74de-4d91-b0a3-dee8336cf9e0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.276870] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Waiting for the task: (returnval){ [ 861.276870] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5246cf58-8fa9-5def-4471-d47b8b27a6bb" [ 861.276870] env[62914]: _type = "Task" [ 861.276870] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.284573] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5246cf58-8fa9-5def-4471-d47b8b27a6bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.354072] env[62914]: INFO nova.compute.manager [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] instance snapshotting [ 861.354392] env[62914]: WARNING nova.compute.manager [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 861.357817] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c56207-02bc-4471-a04e-b8b51eedb70c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.377537] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf347b30-8a72-4b38-a546-c1ae470b899f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.466745] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f71c8436-27ee-4370-ad7d-a38d4f56c816 tempest-MultipleCreateTestJSON-1243532060 tempest-MultipleCreateTestJSON-1243532060-project-member] Lock "9577394d-1c73-4ed1-ba86-e7c246e32719" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.129s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.518061] env[62914]: DEBUG oslo_vmware.api [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352585, 'name': PowerOnVM_Task, 'duration_secs': 0.523455} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.520466] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 861.520703] env[62914]: INFO nova.compute.manager [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Took 8.22 seconds to spawn the instance on the hypervisor. [ 861.520884] env[62914]: DEBUG nova.compute.manager [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.521828] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-affa182e-9cb1-4fd2-8613-54a3206d8769 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.583475] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352589, 'name': Destroy_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.619835] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62d09ed-2201-477b-acfd-c830bb67b1c5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.628276] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b4e4244-55e6-46c3-b4a5-8cd434f66d11 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.660845] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e7b0aa-97ab-4b0f-836f-3ad36a5e9cb5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.668380] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb0b00e-6b30-4da7-af12-61e6dea24c2a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.681545] env[62914]: DEBUG nova.compute.provider_tree [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.736404] env[62914]: DEBUG nova.network.neutron [req-5c0b135d-d783-46c6-95cc-b22fca877835 req-a47608bc-3d53-4809-b0a8-2da0b6e707ad service nova] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Updated VIF entry in instance network info cache for port f1c9a0bc-9d38-477f-a38e-75237f989567. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 861.736798] env[62914]: DEBUG nova.network.neutron [req-5c0b135d-d783-46c6-95cc-b22fca877835 req-a47608bc-3d53-4809-b0a8-2da0b6e707ad service nova] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Updating instance_info_cache with network_info: [{"id": "f1c9a0bc-9d38-477f-a38e-75237f989567", "address": "fa:16:3e:2b:fa:04", "network": {"id": "85e33a61-2e23-4b42-822f-b277a00fd239", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-719568965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b4250bae94c426da87ef1a8d45524d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "21310d90-efbc-45a8-a97f-c4358606530f", "external-id": "nsx-vlan-transportzone-672", "segmentation_id": 672, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1c9a0bc-9d", "ovs_interfaceid": "f1c9a0bc-9d38-477f-a38e-75237f989567", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.787362] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5246cf58-8fa9-5def-4471-d47b8b27a6bb, 'name': SearchDatastore_Task, 'duration_secs': 0.00887} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.787689] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.787930] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 861.788217] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.788425] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.788566] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 861.788829] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5922dc61-ea4a-46c6-abd0-de85d125f9f7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.803156] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 861.803344] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 861.804079] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fed62786-7c88-43e9-b1ca-a0613ac46fc6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.809712] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Waiting for the task: (returnval){ [ 861.809712] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52fb6322-2366-fcfa-c857-b2ce20bd4099" [ 861.809712] env[62914]: _type = "Task" [ 861.809712] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.817359] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52fb6322-2366-fcfa-c857-b2ce20bd4099, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.888378] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Creating Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 861.888720] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-758ea021-508a-434b-8a7f-3de85478fb83 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.896828] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 861.896828] env[62914]: value = "task-1352590" [ 861.896828] env[62914]: _type = "Task" [ 861.896828] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.905901] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352590, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.045631] env[62914]: INFO nova.compute.manager [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Took 26.20 seconds to build instance. [ 862.085106] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352589, 'name': Destroy_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.185214] env[62914]: DEBUG nova.scheduler.client.report [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 862.240664] env[62914]: DEBUG oslo_concurrency.lockutils [req-5c0b135d-d783-46c6-95cc-b22fca877835 req-a47608bc-3d53-4809-b0a8-2da0b6e707ad service nova] Releasing lock "refresh_cache-f80ee46d-9209-4b60-9e5a-f7c8764f4ccb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.321816] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52fb6322-2366-fcfa-c857-b2ce20bd4099, 'name': SearchDatastore_Task, 'duration_secs': 0.035854} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.323052] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d07ab65-411f-4541-9c13-177b022e7c76 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.328410] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Waiting for the task: (returnval){ [ 862.328410] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52cd5269-d836-5cf2-7868-3bcdd19ce288" [ 862.328410] env[62914]: _type = "Task" [ 862.328410] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.338863] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52cd5269-d836-5cf2-7868-3bcdd19ce288, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.406693] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352590, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.549578] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cd9ddfc0-9bc9-400e-8592-9e6ebd849606 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Lock "29680cfb-8df3-4ff7-b6ee-0982577cc9cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.180s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.589101] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352589, 'name': Destroy_Task, 'duration_secs': 1.347976} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.589101] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Destroyed the VM [ 862.589307] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Deleting Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 862.589803] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-19ada707-cab7-4826-b1f6-d232370114ab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.596627] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 862.596627] env[62914]: value = "task-1352591" [ 862.596627] env[62914]: _type = "Task" [ 862.596627] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.605850] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352591, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.692174] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.769s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.694788] env[62914]: DEBUG oslo_concurrency.lockutils [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.140s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.695195] env[62914]: DEBUG nova.objects.instance [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lazy-loading 'resources' on Instance uuid 2274ca27-8e2d-435e-8570-97eb4ae3cc58 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.719662] env[62914]: INFO nova.scheduler.client.report [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Deleted allocations for instance bd3e82ce-62e7-4761-bcd7-d1f835234b97 [ 862.796474] env[62914]: INFO nova.compute.manager [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Rescuing [ 862.796830] env[62914]: DEBUG oslo_concurrency.lockutils [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquiring lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.796978] env[62914]: DEBUG oslo_concurrency.lockutils [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquired lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.797219] env[62914]: DEBUG nova.network.neutron [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 862.842906] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52cd5269-d836-5cf2-7868-3bcdd19ce288, 'name': SearchDatastore_Task, 'duration_secs': 0.010549} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.843422] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.843763] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] f80ee46d-9209-4b60-9e5a-f7c8764f4ccb/f80ee46d-9209-4b60-9e5a-f7c8764f4ccb.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 862.844128] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aa0dd119-db92-4a29-98df-f520bc813401 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.852987] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Waiting for the task: (returnval){ [ 862.852987] env[62914]: value = "task-1352592" [ 862.852987] env[62914]: _type = "Task" [ 862.852987] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.865605] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352592, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.911129] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352590, 'name': CreateSnapshot_Task, 'duration_secs': 0.991034} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.911830] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Created Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 862.912278] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf5d150c-1efa-442d-bc45-489685ca5f61 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.113511] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352591, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.231181] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5323f6a7-e8d4-4c28-a3f0-fcf1168b87c6 tempest-ServersNegativeTestMultiTenantJSON-1030830006 tempest-ServersNegativeTestMultiTenantJSON-1030830006-project-member] Lock "bd3e82ce-62e7-4761-bcd7-d1f835234b97" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.071s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.369317] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352592, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.401156] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac010b12-a641-4516-9ae4-a6eaef2bb442 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.411780] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4563b745-f674-4a87-9899-e652aec3e254 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.452040] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Creating linked-clone VM from snapshot {{(pid=62914) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 863.452376] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5711a27c-edd9-4ea5-814c-61e05e702dd6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.456132] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a9806e-6c45-4c14-82fc-ce7255107165 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.466441] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ffe648-1ae6-4155-93b0-6904825220f3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.470223] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 863.470223] env[62914]: value = "task-1352593" [ 863.470223] env[62914]: _type = "Task" [ 863.470223] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.484588] env[62914]: DEBUG nova.compute.provider_tree [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.491279] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352593, 'name': CloneVM_Task} progress is 11%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.617330] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352591, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.865664] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352592, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526591} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.865938] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] f80ee46d-9209-4b60-9e5a-f7c8764f4ccb/f80ee46d-9209-4b60-9e5a-f7c8764f4ccb.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 863.866509] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.866509] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-96b038dd-883b-4fd9-a480-2d44312acac0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.873902] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Waiting for the task: (returnval){ [ 863.873902] env[62914]: value = "task-1352594" [ 863.873902] env[62914]: _type = "Task" [ 863.873902] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.882689] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352594, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.901577] env[62914]: DEBUG nova.network.neutron [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Updating instance_info_cache with network_info: [{"id": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "address": "fa:16:3e:c9:b5:2c", "network": {"id": "16e49b25-5bce-4cf2-8199-a865e6cd483e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1094957463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "28f67ce59c7f4151b37932bb890c4a5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf4a3fa8-08", "ovs_interfaceid": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.981346] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352593, 'name': CloneVM_Task} progress is 94%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.993511] env[62914]: DEBUG nova.scheduler.client.report [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.112169] env[62914]: DEBUG oslo_vmware.api [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352591, 'name': RemoveSnapshot_Task, 'duration_secs': 1.019241} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.112169] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Deleted Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 864.112169] env[62914]: INFO nova.compute.manager [None req-bf24d198-9453-436f-ae13-e5fbd6481998 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Took 18.92 seconds to snapshot the instance on the hypervisor. [ 864.387025] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352594, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07333} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.387025] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.387025] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3f6526-1764-47ad-81d2-2b2724adfe14 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.404531] env[62914]: DEBUG oslo_concurrency.lockutils [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Releasing lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.417207] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] f80ee46d-9209-4b60-9e5a-f7c8764f4ccb/f80ee46d-9209-4b60-9e5a-f7c8764f4ccb.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.419017] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8bf75fd0-4eba-4c50-b1e4-6dffe62f4a9b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.440728] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Waiting for the task: (returnval){ [ 864.440728] env[62914]: value = "task-1352595" [ 864.440728] env[62914]: _type = "Task" [ 864.440728] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.459111] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352595, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.482641] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352593, 'name': CloneVM_Task} progress is 94%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.501864] env[62914]: DEBUG oslo_concurrency.lockutils [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.807s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.504401] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.552s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.506013] env[62914]: INFO nova.compute.claims [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.538238] env[62914]: INFO nova.scheduler.client.report [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleted allocations for instance 2274ca27-8e2d-435e-8570-97eb4ae3cc58 [ 864.959455] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352595, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.960286] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.960544] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08b5c378-c5a3-4b2e-ab7b-80a827eee9af {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.967927] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 864.967927] env[62914]: value = "task-1352596" [ 864.967927] env[62914]: _type = "Task" [ 864.967927] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.980890] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352596, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.987366] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352593, 'name': CloneVM_Task} progress is 94%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.047044] env[62914]: DEBUG oslo_concurrency.lockutils [None req-131a287f-eab4-41db-b6c7-dc21a551fcbe tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "2274ca27-8e2d-435e-8570-97eb4ae3cc58" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.375s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.297785] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Acquiring lock "47e6f142-bdb2-4397-9b94-d962e68ee6bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.297899] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Lock "47e6f142-bdb2-4397-9b94-d962e68ee6bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.457309] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352595, 'name': ReconfigVM_Task, 'duration_secs': 0.76334} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.457671] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Reconfigured VM instance instance-0000004c to attach disk [datastore2] f80ee46d-9209-4b60-9e5a-f7c8764f4ccb/f80ee46d-9209-4b60-9e5a-f7c8764f4ccb.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.458441] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00374bf1-898a-42ce-9dfd-140352ea78df {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.464842] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Waiting for the task: (returnval){ [ 865.464842] env[62914]: value = "task-1352597" [ 865.464842] env[62914]: _type = "Task" [ 865.464842] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.487076] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352596, 'name': PowerOffVM_Task, 'duration_secs': 0.307844} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.487453] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352597, 'name': Rename_Task} progress is 10%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.488134] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 865.488958] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b352c0be-f6f9-4b04-a147-5cc569b007eb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.495033] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352593, 'name': CloneVM_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.515194] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9c950a-307b-4e9f-b929-4c2a7da6368f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.556926] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 865.557285] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa2c06e4-5c56-4560-8312-b59d1162175c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.567668] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 865.567668] env[62914]: value = "task-1352598" [ 865.567668] env[62914]: _type = "Task" [ 865.567668] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.580881] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] VM already powered off {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 865.580881] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.581580] env[62914]: DEBUG oslo_concurrency.lockutils [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.581580] env[62914]: DEBUG oslo_concurrency.lockutils [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.581580] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.581802] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e869a7d-91bf-4e98-ab80-40652f5901df {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.591963] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.591963] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 865.592402] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29956f26-c131-40e9-b5ae-2b32b0821c27 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.601258] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 865.601258] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]523748bc-840d-97ee-0448-c9dac71f15b6" [ 865.601258] env[62914]: _type = "Task" [ 865.601258] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.614033] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]523748bc-840d-97ee-0448-c9dac71f15b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.744918] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102f6168-dcef-47b3-a099-696ecbad50f5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.752516] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c21a18-e6ef-45bc-835a-e81166b1d186 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.787507] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c800069-002c-4163-a996-3a8388c7785a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.795948] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-345ff6a4-2f27-45c1-b21a-f57e8d07da27 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.800723] env[62914]: DEBUG nova.compute.manager [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 865.816919] env[62914]: DEBUG nova.compute.provider_tree [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.982694] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352597, 'name': Rename_Task, 'duration_secs': 0.152818} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.986867] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 865.986867] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46c8834d-5804-4e6b-b4e4-51a894afe32c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.000143] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352593, 'name': CloneVM_Task, 'duration_secs': 2.056829} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.000574] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Waiting for the task: (returnval){ [ 866.000574] env[62914]: value = "task-1352599" [ 866.000574] env[62914]: _type = "Task" [ 866.000574] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.001035] env[62914]: INFO nova.virt.vmwareapi.vmops [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Created linked-clone VM from snapshot [ 866.003395] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92787f89-9524-4b51-aa47-6ae8e4dce6fd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.018029] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Uploading image 0b52b7b9-4343-488f-b3ec-34bf2b5cff88 {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 866.020476] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352599, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.047020] env[62914]: DEBUG oslo_vmware.rw_handles [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 866.047020] env[62914]: value = "vm-288210" [ 866.047020] env[62914]: _type = "VirtualMachine" [ 866.047020] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 866.047020] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1f8ab532-ee2a-48a0-88e3-13ccb495088e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.052052] env[62914]: DEBUG oslo_vmware.rw_handles [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lease: (returnval){ [ 866.052052] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]527caf55-c62a-4146-3758-99bfe6725a20" [ 866.052052] env[62914]: _type = "HttpNfcLease" [ 866.052052] env[62914]: } obtained for exporting VM: (result){ [ 866.052052] env[62914]: value = "vm-288210" [ 866.052052] env[62914]: _type = "VirtualMachine" [ 866.052052] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 866.052387] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the lease: (returnval){ [ 866.052387] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]527caf55-c62a-4146-3758-99bfe6725a20" [ 866.052387] env[62914]: _type = "HttpNfcLease" [ 866.052387] env[62914]: } to be ready. {{(pid=62914) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 866.058412] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 866.058412] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]527caf55-c62a-4146-3758-99bfe6725a20" [ 866.058412] env[62914]: _type = "HttpNfcLease" [ 866.058412] env[62914]: } is initializing. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 866.110994] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]523748bc-840d-97ee-0448-c9dac71f15b6, 'name': SearchDatastore_Task, 'duration_secs': 0.00983} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.111958] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28e5fc42-31a6-406a-b0fc-394c088bb8be {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.117159] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 866.117159] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5285ad1e-b571-eb7d-643e-a512fe1fcf9b" [ 866.117159] env[62914]: _type = "Task" [ 866.117159] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.125964] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5285ad1e-b571-eb7d-643e-a512fe1fcf9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.145665] env[62914]: DEBUG oslo_concurrency.lockutils [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Acquiring lock "9fab500c-344c-46cd-b060-c8f1aa0f1cba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.145939] env[62914]: DEBUG oslo_concurrency.lockutils [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Lock "9fab500c-344c-46cd-b060-c8f1aa0f1cba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.146172] env[62914]: DEBUG oslo_concurrency.lockutils [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Acquiring lock "9fab500c-344c-46cd-b060-c8f1aa0f1cba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.146358] env[62914]: DEBUG oslo_concurrency.lockutils [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Lock "9fab500c-344c-46cd-b060-c8f1aa0f1cba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.146527] env[62914]: DEBUG oslo_concurrency.lockutils [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Lock "9fab500c-344c-46cd-b060-c8f1aa0f1cba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.152304] env[62914]: INFO nova.compute.manager [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Terminating instance [ 866.153944] env[62914]: DEBUG nova.compute.manager [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 866.154157] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 866.154981] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e86c2a4-a5f6-4c65-9ad0-b701d3feee9b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.164185] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.165132] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1dbb69a4-2e43-400c-9e9d-d940e61bf9da {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.170382] env[62914]: DEBUG oslo_vmware.api [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Waiting for the task: (returnval){ [ 866.170382] env[62914]: value = "task-1352601" [ 866.170382] env[62914]: _type = "Task" [ 866.170382] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.178799] env[62914]: DEBUG oslo_vmware.api [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352601, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.320499] env[62914]: DEBUG nova.scheduler.client.report [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.338972] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.416891] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "c9b94435-6546-464a-b3ba-c9e685157059" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.416983] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "c9b94435-6546-464a-b3ba-c9e685157059" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.512567] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352599, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.560700] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 866.560700] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]527caf55-c62a-4146-3758-99bfe6725a20" [ 866.560700] env[62914]: _type = "HttpNfcLease" [ 866.560700] env[62914]: } is ready. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 866.561042] env[62914]: DEBUG oslo_vmware.rw_handles [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 866.561042] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]527caf55-c62a-4146-3758-99bfe6725a20" [ 866.561042] env[62914]: _type = "HttpNfcLease" [ 866.561042] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 866.561823] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a59010-8a7a-457e-85aa-a84fc699586a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.570305] env[62914]: DEBUG oslo_vmware.rw_handles [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527ad015-0af4-f987-187d-d7a213c98f1d/disk-0.vmdk from lease info. {{(pid=62914) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 866.570480] env[62914]: DEBUG oslo_vmware.rw_handles [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527ad015-0af4-f987-187d-d7a213c98f1d/disk-0.vmdk for reading. {{(pid=62914) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 866.642285] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5285ad1e-b571-eb7d-643e-a512fe1fcf9b, 'name': SearchDatastore_Task, 'duration_secs': 0.011437} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.642620] env[62914]: DEBUG oslo_concurrency.lockutils [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.642863] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 29680cfb-8df3-4ff7-b6ee-0982577cc9cf/6c21e11e-da68-48a7-9929-737212ce98cf-rescue.vmdk. {{(pid=62914) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 866.643354] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-555d2e6e-b51b-41e5-975d-0b8eb4c30abd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.650500] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 866.650500] env[62914]: value = "task-1352602" [ 866.650500] env[62914]: _type = "Task" [ 866.650500] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.659194] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352602, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.678151] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f973437f-4063-46f2-881c-e8c3487d4737 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.684493] env[62914]: DEBUG oslo_vmware.api [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352601, 'name': PowerOffVM_Task, 'duration_secs': 0.241682} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.684631] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 866.684816] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 866.685077] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e4f5fb4f-d5ee-49c1-951d-7336e00b637f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.761381] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 866.761381] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 866.761381] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Deleting the datastore file [datastore2] 9fab500c-344c-46cd-b060-c8f1aa0f1cba {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 866.761661] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2ba05eb3-b3da-4c5c-a4fb-9e45512cf13b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.767863] env[62914]: DEBUG oslo_vmware.api [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Waiting for the task: (returnval){ [ 866.767863] env[62914]: value = "task-1352604" [ 866.767863] env[62914]: _type = "Task" [ 866.767863] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.777406] env[62914]: DEBUG oslo_vmware.api [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352604, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.826698] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.827387] env[62914]: DEBUG nova.compute.manager [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 866.834378] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 8.415s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.919962] env[62914]: DEBUG nova.compute.manager [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 867.015870] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352599, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.163109] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352602, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.234291] env[62914]: DEBUG nova.compute.manager [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.235095] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c29da03e-852d-4797-b399-cb15782f9e67 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.283199] env[62914]: DEBUG oslo_vmware.api [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Task: {'id': task-1352604, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.453056} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.284523] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.284885] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.285450] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.285742] env[62914]: INFO nova.compute.manager [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Took 1.13 seconds to destroy the instance on the hypervisor. [ 867.286114] env[62914]: DEBUG oslo.service.loopingcall [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.286443] env[62914]: DEBUG nova.compute.manager [-] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.286677] env[62914]: DEBUG nova.network.neutron [-] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 867.336622] env[62914]: DEBUG nova.compute.utils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 867.342917] env[62914]: DEBUG nova.compute.manager [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 867.343304] env[62914]: DEBUG nova.network.neutron [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 867.351672] env[62914]: INFO nova.compute.claims [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.448014] env[62914]: DEBUG nova.policy [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2c21c59e39b94deea6c3701fe971198a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1dcda44f0d6a455bb17966e7c5061190', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 867.453639] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.516360] env[62914]: DEBUG oslo_vmware.api [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352599, 'name': PowerOnVM_Task, 'duration_secs': 1.281035} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.517080] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.517080] env[62914]: INFO nova.compute.manager [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Took 9.05 seconds to spawn the instance on the hypervisor. [ 867.521838] env[62914]: DEBUG nova.compute.manager [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.521838] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ba6c85-9d63-4950-b2c7-df1060fbb6ad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.664668] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352602, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518503} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.664982] env[62914]: INFO nova.virt.vmwareapi.ds_util [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 29680cfb-8df3-4ff7-b6ee-0982577cc9cf/6c21e11e-da68-48a7-9929-737212ce98cf-rescue.vmdk. [ 867.665794] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b6977d3-6877-4c9d-b9aa-bb0dea68978c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.696009] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 29680cfb-8df3-4ff7-b6ee-0982577cc9cf/6c21e11e-da68-48a7-9929-737212ce98cf-rescue.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.696498] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-469c1366-177d-4a1e-96b1-047e8eca44dc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.716903] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 867.716903] env[62914]: value = "task-1352605" [ 867.716903] env[62914]: _type = "Task" [ 867.716903] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.726636] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352605, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.747143] env[62914]: INFO nova.compute.manager [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] instance snapshotting [ 867.753471] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839b1664-cae8-489f-bdcf-034b31dd38b2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.771833] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a401b571-37f5-408b-a851-075411ee83bf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.827073] env[62914]: DEBUG nova.network.neutron [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Successfully created port: 02c6f8b8-aaaa-4b7e-979c-43ad85975569 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 867.858909] env[62914]: DEBUG nova.compute.manager [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 867.865195] env[62914]: INFO nova.compute.resource_tracker [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating resource usage from migration 19c9417f-8ea9-4e58-b81e-a66228590190 [ 868.042659] env[62914]: INFO nova.compute.manager [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Took 26.61 seconds to build instance. [ 868.090411] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2722c2f5-4063-40f1-95e9-fd6ae9fdacf0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.101816] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ac2e81-f1a2-48ba-8f73-458e8e9de78b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.144377] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a754ec1d-15da-441c-a555-a81e9c6f7512 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.154022] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9283de-71de-4bc1-816d-5d90a5fd80bf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.170694] env[62914]: DEBUG nova.compute.provider_tree [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.228409] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352605, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.237364] env[62914]: DEBUG nova.compute.manager [req-1fb7e2b9-a79b-408d-ae35-ac37b196b679 req-333a3322-0a1d-40e1-a5ae-5d384c0ef73f service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Received event network-vif-deleted-299c1132-e4a5-4803-8e8d-1478b5928463 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.237589] env[62914]: INFO nova.compute.manager [req-1fb7e2b9-a79b-408d-ae35-ac37b196b679 req-333a3322-0a1d-40e1-a5ae-5d384c0ef73f service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Neutron deleted interface 299c1132-e4a5-4803-8e8d-1478b5928463; detaching it from the instance and deleting it from the info cache [ 868.237868] env[62914]: DEBUG nova.network.neutron [req-1fb7e2b9-a79b-408d-ae35-ac37b196b679 req-333a3322-0a1d-40e1-a5ae-5d384c0ef73f service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.285323] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Creating Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 868.285637] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a7573dcb-3a5e-4126-b371-dde15c8d0f8d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.292969] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 868.292969] env[62914]: value = "task-1352606" [ 868.292969] env[62914]: _type = "Task" [ 868.292969] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.302668] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352606, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.401171] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Acquiring lock "f80ee46d-9209-4b60-9e5a-f7c8764f4ccb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.545015] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d49e0702-1169-496a-bd2d-67476c3cb0fc tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Lock "f80ee46d-9209-4b60-9e5a-f7c8764f4ccb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.905s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.545338] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Lock "f80ee46d-9209-4b60-9e5a-f7c8764f4ccb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.144s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.545573] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Acquiring lock "f80ee46d-9209-4b60-9e5a-f7c8764f4ccb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.545825] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Lock "f80ee46d-9209-4b60-9e5a-f7c8764f4ccb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.545957] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Lock "f80ee46d-9209-4b60-9e5a-f7c8764f4ccb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.548250] env[62914]: INFO nova.compute.manager [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Terminating instance [ 868.550486] env[62914]: DEBUG nova.compute.manager [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 868.550486] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 868.551390] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df961b96-425c-4778-bcf1-b58bf39014e0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.559131] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 868.559416] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6379f4a2-13ba-4a71-93dc-37113ef970e6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.566074] env[62914]: DEBUG oslo_vmware.api [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Waiting for the task: (returnval){ [ 868.566074] env[62914]: value = "task-1352607" [ 868.566074] env[62914]: _type = "Task" [ 868.566074] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.574617] env[62914]: DEBUG oslo_vmware.api [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352607, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.675545] env[62914]: DEBUG nova.scheduler.client.report [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.713484] env[62914]: DEBUG nova.network.neutron [-] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.742355] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352605, 'name': ReconfigVM_Task, 'duration_secs': 0.667425} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.742910] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ebf4991-8edb-48f9-9d81-fdbb105a05ab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.746119] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 29680cfb-8df3-4ff7-b6ee-0982577cc9cf/6c21e11e-da68-48a7-9929-737212ce98cf-rescue.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.749295] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86fe087f-2161-4dea-b0e8-622ed214e0de {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.797777] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ed6c598-4221-4812-b2e8-53ca7c3034e6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.825926] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-564ddfcd-234f-4866-b346-ad69c37a11f0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.851063] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352606, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.853811] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 868.853811] env[62914]: value = "task-1352608" [ 868.853811] env[62914]: _type = "Task" [ 868.853811] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.868031] env[62914]: DEBUG nova.compute.manager [req-1fb7e2b9-a79b-408d-ae35-ac37b196b679 req-333a3322-0a1d-40e1-a5ae-5d384c0ef73f service nova] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Detach interface failed, port_id=299c1132-e4a5-4803-8e8d-1478b5928463, reason: Instance 9fab500c-344c-46cd-b060-c8f1aa0f1cba could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 868.881235] env[62914]: DEBUG nova.compute.manager [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 868.884294] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352608, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.915215] env[62914]: DEBUG nova.virt.hardware [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 868.915215] env[62914]: DEBUG nova.virt.hardware [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 868.915215] env[62914]: DEBUG nova.virt.hardware [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.915447] env[62914]: DEBUG nova.virt.hardware [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 868.915447] env[62914]: DEBUG nova.virt.hardware [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.915447] env[62914]: DEBUG nova.virt.hardware [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 868.915596] env[62914]: DEBUG nova.virt.hardware [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 868.920451] env[62914]: DEBUG nova.virt.hardware [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 868.920451] env[62914]: DEBUG nova.virt.hardware [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 868.920451] env[62914]: DEBUG nova.virt.hardware [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 868.920451] env[62914]: DEBUG nova.virt.hardware [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.920451] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7190a70-b1b0-4887-92ea-f2e87142490f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.930776] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2ea7a9-e4b3-411f-8497-232a65117899 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.080701] env[62914]: DEBUG oslo_vmware.api [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352607, 'name': PowerOffVM_Task, 'duration_secs': 0.196876} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.082031] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 869.082031] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 869.082031] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8c2dae28-627f-4e22-9c02-2b2accfebf4e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.154618] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 869.154954] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 869.155326] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Deleting the datastore file [datastore2] f80ee46d-9209-4b60-9e5a-f7c8764f4ccb {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.155614] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-677bebed-7b6e-455f-8ebb-f873fbc75ffb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.162478] env[62914]: DEBUG oslo_vmware.api [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Waiting for the task: (returnval){ [ 869.162478] env[62914]: value = "task-1352610" [ 869.162478] env[62914]: _type = "Task" [ 869.162478] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.171396] env[62914]: DEBUG oslo_vmware.api [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352610, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.182221] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.350s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.182499] env[62914]: INFO nova.compute.manager [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Migrating [ 869.182706] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.182848] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.184568] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 8.176s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.215858] env[62914]: INFO nova.compute.manager [-] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Took 1.93 seconds to deallocate network for instance. [ 869.310137] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352606, 'name': CreateSnapshot_Task, 'duration_secs': 0.912861} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.310431] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Created Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 869.311227] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9b0701-9bc0-4d87-a943-ab04c7af9ea5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.376707] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352608, 'name': ReconfigVM_Task, 'duration_secs': 0.240267} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.377020] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 869.377294] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09e3ca9f-743e-49cd-a5fb-284567f8904a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.384995] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 869.384995] env[62914]: value = "task-1352611" [ 869.384995] env[62914]: _type = "Task" [ 869.384995] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.393785] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352611, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.404765] env[62914]: DEBUG nova.compute.manager [req-d118ef01-0b86-426f-8764-1145cd0191b6 req-6da5479f-6d56-419c-8e77-e31a7cc2a0b9 service nova] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Received event network-vif-plugged-02c6f8b8-aaaa-4b7e-979c-43ad85975569 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.404971] env[62914]: DEBUG oslo_concurrency.lockutils [req-d118ef01-0b86-426f-8764-1145cd0191b6 req-6da5479f-6d56-419c-8e77-e31a7cc2a0b9 service nova] Acquiring lock "22de53a6-096b-4e8e-af91-f78588ea4c60-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.405240] env[62914]: DEBUG oslo_concurrency.lockutils [req-d118ef01-0b86-426f-8764-1145cd0191b6 req-6da5479f-6d56-419c-8e77-e31a7cc2a0b9 service nova] Lock "22de53a6-096b-4e8e-af91-f78588ea4c60-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.405416] env[62914]: DEBUG oslo_concurrency.lockutils [req-d118ef01-0b86-426f-8764-1145cd0191b6 req-6da5479f-6d56-419c-8e77-e31a7cc2a0b9 service nova] Lock "22de53a6-096b-4e8e-af91-f78588ea4c60-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.405688] env[62914]: DEBUG nova.compute.manager [req-d118ef01-0b86-426f-8764-1145cd0191b6 req-6da5479f-6d56-419c-8e77-e31a7cc2a0b9 service nova] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] No waiting events found dispatching network-vif-plugged-02c6f8b8-aaaa-4b7e-979c-43ad85975569 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 869.405757] env[62914]: WARNING nova.compute.manager [req-d118ef01-0b86-426f-8764-1145cd0191b6 req-6da5479f-6d56-419c-8e77-e31a7cc2a0b9 service nova] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Received unexpected event network-vif-plugged-02c6f8b8-aaaa-4b7e-979c-43ad85975569 for instance with vm_state building and task_state spawning. [ 869.486720] env[62914]: DEBUG nova.network.neutron [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Successfully updated port: 02c6f8b8-aaaa-4b7e-979c-43ad85975569 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 869.548695] env[62914]: DEBUG oslo_concurrency.lockutils [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "interface-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.549053] env[62914]: DEBUG oslo_concurrency.lockutils [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.549294] env[62914]: DEBUG nova.objects.instance [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'flavor' on Instance uuid 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.673603] env[62914]: DEBUG oslo_vmware.api [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Task: {'id': task-1352610, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.448417} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.673864] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.674071] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 869.674258] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 869.674434] env[62914]: INFO nova.compute.manager [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Took 1.12 seconds to destroy the instance on the hypervisor. [ 869.674679] env[62914]: DEBUG oslo.service.loopingcall [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.674872] env[62914]: DEBUG nova.compute.manager [-] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 869.675108] env[62914]: DEBUG nova.network.neutron [-] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 869.686352] env[62914]: INFO nova.compute.rpcapi [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 869.686959] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.723062] env[62914]: DEBUG oslo_concurrency.lockutils [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.829875] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Creating linked-clone VM from snapshot {{(pid=62914) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 869.830368] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7960d39f-7353-474d-85ca-1dfaad33fec7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.844281] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 869.844281] env[62914]: value = "task-1352612" [ 869.844281] env[62914]: _type = "Task" [ 869.844281] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.856799] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352612, 'name': CloneVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.897418] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352611, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.990792] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "refresh_cache-22de53a6-096b-4e8e-af91-f78588ea4c60" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.990897] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "refresh_cache-22de53a6-096b-4e8e-af91-f78588ea4c60" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.991871] env[62914]: DEBUG nova.network.neutron [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 870.149462] env[62914]: DEBUG nova.objects.instance [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'pci_requests' on Instance uuid 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.207842] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Applying migration context for instance 4bf76ce0-b5d2-4184-888c-d0ef39878356 as it has an incoming, in-progress migration 19c9417f-8ea9-4e58-b81e-a66228590190. Migration status is pre-migrating {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 870.209456] env[62914]: INFO nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating resource usage from migration 19c9417f-8ea9-4e58-b81e-a66228590190 [ 870.212245] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.212488] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.212875] env[62914]: DEBUG nova.network.neutron [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 870.238049] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 9fab500c-344c-46cd-b060-c8f1aa0f1cba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 870.238049] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 870.238049] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 93d420a1-6d8f-4919-a42f-55aebab853ae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 870.238049] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 1fe15423-f8be-4763-b55e-2cbb383ff01d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 870.238278] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance c7cf71d7-b823-471f-949f-9ef2b539d972 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 870.238278] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 43ee8948-8805-4d48-bd45-e93d2e2eb05d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 870.238278] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 29680cfb-8df3-4ff7-b6ee-0982577cc9cf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 870.238278] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance f80ee46d-9209-4b60-9e5a-f7c8764f4ccb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 870.238389] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 22de53a6-096b-4e8e-af91-f78588ea4c60 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 870.238389] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Migration 19c9417f-8ea9-4e58-b81e-a66228590190 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 870.238544] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 4bf76ce0-b5d2-4184-888c-d0ef39878356 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 870.354701] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352612, 'name': CloneVM_Task} progress is 94%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.396787] env[62914]: DEBUG oslo_vmware.api [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352611, 'name': PowerOnVM_Task, 'duration_secs': 0.843506} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.397104] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 870.400201] env[62914]: DEBUG nova.compute.manager [None req-50380ea1-f47f-4793-acc3-2c18b2476f31 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.401263] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9332ad4-b95f-47f1-9fcb-803144d8f982 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.473121] env[62914]: DEBUG nova.network.neutron [-] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.521217] env[62914]: DEBUG nova.network.neutron [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 870.652574] env[62914]: DEBUG nova.objects.base [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Object Instance<0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07> lazy-loaded attributes: flavor,pci_requests {{(pid=62914) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 870.652901] env[62914]: DEBUG nova.network.neutron [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 870.663190] env[62914]: DEBUG nova.network.neutron [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Updating instance_info_cache with network_info: [{"id": "02c6f8b8-aaaa-4b7e-979c-43ad85975569", "address": "fa:16:3e:23:52:28", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02c6f8b8-aa", "ovs_interfaceid": "02c6f8b8-aaaa-4b7e-979c-43ad85975569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.693483] env[62914]: DEBUG nova.policy [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80089586cfda4cd2bdfc24d9f556239c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09d51090ac5d41fc966c7e132f288824', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 870.741913] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 47e6f142-bdb2-4397-9b94-d962e68ee6bb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 870.862735] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352612, 'name': CloneVM_Task} progress is 94%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.976316] env[62914]: INFO nova.compute.manager [-] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Took 1.30 seconds to deallocate network for instance. [ 870.994772] env[62914]: DEBUG nova.network.neutron [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Successfully created port: 69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 871.005541] env[62914]: DEBUG nova.network.neutron [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating instance_info_cache with network_info: [{"id": "189c8622-547d-4da9-94da-564df3bc382d", "address": "fa:16:3e:b6:1a:f0", "network": {"id": "21b6ff73-feab-478c-848a-313ea0acbfc2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1273337660-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "619be7b3bf1445b68ba7adefc98f8782", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap189c8622-54", "ovs_interfaceid": "189c8622-547d-4da9-94da-564df3bc382d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.168340] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "refresh_cache-22de53a6-096b-4e8e-af91-f78588ea4c60" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.168340] env[62914]: DEBUG nova.compute.manager [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Instance network_info: |[{"id": "02c6f8b8-aaaa-4b7e-979c-43ad85975569", "address": "fa:16:3e:23:52:28", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02c6f8b8-aa", "ovs_interfaceid": "02c6f8b8-aaaa-4b7e-979c-43ad85975569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 871.168657] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:52:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a8c8175-1197-4f12-baac-ef6aba95f585', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '02c6f8b8-aaaa-4b7e-979c-43ad85975569', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 871.177383] env[62914]: DEBUG oslo.service.loopingcall [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 871.177383] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 871.177784] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-51782e4a-9bff-4739-be2f-c06328d504b0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.199499] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 871.199499] env[62914]: value = "task-1352613" [ 871.199499] env[62914]: _type = "Task" [ 871.199499] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.208755] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352613, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.244817] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance c9b94435-6546-464a-b3ba-c9e685157059 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 871.245111] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 871.245262] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 871.360191] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352612, 'name': CloneVM_Task, 'duration_secs': 1.392718} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.362892] env[62914]: INFO nova.virt.vmwareapi.vmops [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Created linked-clone VM from snapshot [ 871.362892] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84339346-ac5c-4d51-8641-c8ceb94b16e3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.372220] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Uploading image 8e275538-804d-43f7-8258-35fc043e5a94 {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 871.404062] env[62914]: DEBUG oslo_vmware.rw_handles [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 871.404062] env[62914]: value = "vm-288212" [ 871.404062] env[62914]: _type = "VirtualMachine" [ 871.404062] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 871.404396] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8e2b6eb3-f0e5-437c-a8a8-3bc85c9b8453 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.415787] env[62914]: DEBUG oslo_vmware.rw_handles [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lease: (returnval){ [ 871.415787] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]521fe1d5-578f-e00a-3835-c1539d72a69f" [ 871.415787] env[62914]: _type = "HttpNfcLease" [ 871.415787] env[62914]: } obtained for exporting VM: (result){ [ 871.415787] env[62914]: value = "vm-288212" [ 871.415787] env[62914]: _type = "VirtualMachine" [ 871.415787] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 871.416173] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the lease: (returnval){ [ 871.416173] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]521fe1d5-578f-e00a-3835-c1539d72a69f" [ 871.416173] env[62914]: _type = "HttpNfcLease" [ 871.416173] env[62914]: } to be ready. {{(pid=62914) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 871.427944] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 871.427944] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]521fe1d5-578f-e00a-3835-c1539d72a69f" [ 871.427944] env[62914]: _type = "HttpNfcLease" [ 871.427944] env[62914]: } is initializing. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 871.448906] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd435fbc-b99e-4960-a488-e43b286ca972 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.457014] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2686118c-113e-4552-8a30-1e77be83b68b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.487945] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.488967] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7434ce19-97ef-41e3-b343-0020b8485802 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.496784] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca0ec7c5-8684-4eca-9d5a-7ee516a098e4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.510737] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.513015] env[62914]: DEBUG nova.compute.provider_tree [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.712050] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352613, 'name': CreateVM_Task, 'duration_secs': 0.359217} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.712050] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 871.712050] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.712050] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.712050] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 871.712050] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d33f26b9-1ca6-495b-b769-7a8669f22103 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.716501] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 871.716501] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c29897-8791-1eaa-81d4-bf7801221e9d" [ 871.716501] env[62914]: _type = "Task" [ 871.716501] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.726176] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c29897-8791-1eaa-81d4-bf7801221e9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.843362] env[62914]: DEBUG nova.compute.manager [req-922baa38-e448-4e78-8aeb-13144897c3ba req-7d815f10-581e-40ba-878a-0e116220f74b service nova] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Received event network-changed-02c6f8b8-aaaa-4b7e-979c-43ad85975569 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.843449] env[62914]: DEBUG nova.compute.manager [req-922baa38-e448-4e78-8aeb-13144897c3ba req-7d815f10-581e-40ba-878a-0e116220f74b service nova] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Refreshing instance network info cache due to event network-changed-02c6f8b8-aaaa-4b7e-979c-43ad85975569. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 871.843662] env[62914]: DEBUG oslo_concurrency.lockutils [req-922baa38-e448-4e78-8aeb-13144897c3ba req-7d815f10-581e-40ba-878a-0e116220f74b service nova] Acquiring lock "refresh_cache-22de53a6-096b-4e8e-af91-f78588ea4c60" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.843808] env[62914]: DEBUG oslo_concurrency.lockutils [req-922baa38-e448-4e78-8aeb-13144897c3ba req-7d815f10-581e-40ba-878a-0e116220f74b service nova] Acquired lock "refresh_cache-22de53a6-096b-4e8e-af91-f78588ea4c60" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.843968] env[62914]: DEBUG nova.network.neutron [req-922baa38-e448-4e78-8aeb-13144897c3ba req-7d815f10-581e-40ba-878a-0e116220f74b service nova] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Refreshing network info cache for port 02c6f8b8-aaaa-4b7e-979c-43ad85975569 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 871.924948] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 871.924948] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]521fe1d5-578f-e00a-3835-c1539d72a69f" [ 871.924948] env[62914]: _type = "HttpNfcLease" [ 871.924948] env[62914]: } is ready. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 871.925307] env[62914]: DEBUG oslo_vmware.rw_handles [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 871.925307] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]521fe1d5-578f-e00a-3835-c1539d72a69f" [ 871.925307] env[62914]: _type = "HttpNfcLease" [ 871.925307] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 871.926026] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b89f33-3333-4842-9b54-9241a478a23f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.933378] env[62914]: DEBUG oslo_vmware.rw_handles [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5298fab9-1ad3-f8fe-f920-0b2ae67e3957/disk-0.vmdk from lease info. {{(pid=62914) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 871.933558] env[62914]: DEBUG oslo_vmware.rw_handles [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5298fab9-1ad3-f8fe-f920-0b2ae67e3957/disk-0.vmdk for reading. {{(pid=62914) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 872.018231] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.044977] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3afdb256-55aa-4bd4-bf0c-57329683bc8a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.229527] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c29897-8791-1eaa-81d4-bf7801221e9d, 'name': SearchDatastore_Task, 'duration_secs': 0.012598} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.230086] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.230559] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 872.231019] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.231360] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.231955] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 872.232340] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-840567f1-880b-44f9-b587-20a3bb71fd3f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.244213] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 872.244213] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 872.244213] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa60903a-862c-473c-a454-6d8bf34c2d79 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.248982] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 872.248982] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52711f79-467f-62d2-3e8f-7b5e577df505" [ 872.248982] env[62914]: _type = "Task" [ 872.248982] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.260228] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52711f79-467f-62d2-3e8f-7b5e577df505, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.525366] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 872.525579] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.341s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.526812] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.189s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.531190] env[62914]: INFO nova.compute.claims [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.533312] env[62914]: DEBUG nova.network.neutron [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Successfully updated port: 69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 872.701693] env[62914]: DEBUG nova.network.neutron [req-922baa38-e448-4e78-8aeb-13144897c3ba req-7d815f10-581e-40ba-878a-0e116220f74b service nova] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Updated VIF entry in instance network info cache for port 02c6f8b8-aaaa-4b7e-979c-43ad85975569. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 872.702336] env[62914]: DEBUG nova.network.neutron [req-922baa38-e448-4e78-8aeb-13144897c3ba req-7d815f10-581e-40ba-878a-0e116220f74b service nova] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Updating instance_info_cache with network_info: [{"id": "02c6f8b8-aaaa-4b7e-979c-43ad85975569", "address": "fa:16:3e:23:52:28", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02c6f8b8-aa", "ovs_interfaceid": "02c6f8b8-aaaa-4b7e-979c-43ad85975569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.760287] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52711f79-467f-62d2-3e8f-7b5e577df505, 'name': SearchDatastore_Task, 'duration_secs': 0.010736} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.761403] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3817d21e-20c7-462c-9054-2e4fc76c0cec {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.766870] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 872.766870] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d5f4fb-58fb-1983-ef48-55dfd619dff3" [ 872.766870] env[62914]: _type = "Task" [ 872.766870] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.775168] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d5f4fb-58fb-1983-ef48-55dfd619dff3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.031038] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f90a5e47-0c09-48f8-bac2-ff81b318f957 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.038751] env[62914]: DEBUG oslo_concurrency.lockutils [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.039015] env[62914]: DEBUG oslo_concurrency.lockutils [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.039086] env[62914]: DEBUG nova.network.neutron [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 873.058390] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating instance '4bf76ce0-b5d2-4184-888c-d0ef39878356' progress to 0 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 873.207675] env[62914]: DEBUG oslo_concurrency.lockutils [req-922baa38-e448-4e78-8aeb-13144897c3ba req-7d815f10-581e-40ba-878a-0e116220f74b service nova] Releasing lock "refresh_cache-22de53a6-096b-4e8e-af91-f78588ea4c60" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.208301] env[62914]: DEBUG nova.compute.manager [req-922baa38-e448-4e78-8aeb-13144897c3ba req-7d815f10-581e-40ba-878a-0e116220f74b service nova] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Received event network-vif-deleted-f1c9a0bc-9d38-477f-a38e-75237f989567 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.210342] env[62914]: DEBUG nova.compute.manager [req-1b2ffe9f-f0f1-447a-8000-e7dd39734bba req-5c3a6cb1-2722-4643-8c55-1a9484391f97 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Received event network-vif-plugged-69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.210476] env[62914]: DEBUG oslo_concurrency.lockutils [req-1b2ffe9f-f0f1-447a-8000-e7dd39734bba req-5c3a6cb1-2722-4643-8c55-1a9484391f97 service nova] Acquiring lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.210737] env[62914]: DEBUG oslo_concurrency.lockutils [req-1b2ffe9f-f0f1-447a-8000-e7dd39734bba req-5c3a6cb1-2722-4643-8c55-1a9484391f97 service nova] Lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.210996] env[62914]: DEBUG oslo_concurrency.lockutils [req-1b2ffe9f-f0f1-447a-8000-e7dd39734bba req-5c3a6cb1-2722-4643-8c55-1a9484391f97 service nova] Lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.211278] env[62914]: DEBUG nova.compute.manager [req-1b2ffe9f-f0f1-447a-8000-e7dd39734bba req-5c3a6cb1-2722-4643-8c55-1a9484391f97 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] No waiting events found dispatching network-vif-plugged-69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 873.211472] env[62914]: WARNING nova.compute.manager [req-1b2ffe9f-f0f1-447a-8000-e7dd39734bba req-5c3a6cb1-2722-4643-8c55-1a9484391f97 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Received unexpected event network-vif-plugged-69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6 for instance with vm_state active and task_state None. [ 873.278577] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d5f4fb-58fb-1983-ef48-55dfd619dff3, 'name': SearchDatastore_Task, 'duration_secs': 0.013218} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.278805] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.279037] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 22de53a6-096b-4e8e-af91-f78588ea4c60/22de53a6-096b-4e8e-af91-f78588ea4c60.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 873.279450] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1bf76f8-c1c4-42c3-9607-54f6e717de3e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.286064] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 873.286064] env[62914]: value = "task-1352615" [ 873.286064] env[62914]: _type = "Task" [ 873.286064] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.294136] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352615, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.566510] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 873.569579] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0b550d68-1e0e-4ae4-bf65-72c1610254ba {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.577258] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 873.577258] env[62914]: value = "task-1352616" [ 873.577258] env[62914]: _type = "Task" [ 873.577258] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.581821] env[62914]: WARNING nova.network.neutron [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] f59e1bed-468c-492e-a501-34925e946a66 already exists in list: networks containing: ['f59e1bed-468c-492e-a501-34925e946a66']. ignoring it [ 873.591522] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352616, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.758091] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0a8c10-b114-4246-a979-b84356885b8d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.767501] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e2aec5-b820-4eba-94a3-4535a3bd9b7d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.814357] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b07b7c8-7bd5-4621-b400-3324842a4769 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.823631] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352615, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.827689] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1e8ed1-f8f2-4240-bf02-66265f901de5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.844298] env[62914]: DEBUG nova.compute.provider_tree [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.962948] env[62914]: DEBUG nova.network.neutron [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Updating instance_info_cache with network_info: [{"id": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "address": "fa:16:3e:7a:f0:fa", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb12a55b6-c3", "ovs_interfaceid": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6", "address": "fa:16:3e:c1:ef:cc", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69f9fa9f-a0", "ovs_interfaceid": "69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.089313] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352616, 'name': PowerOffVM_Task, 'duration_secs': 0.244038} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.089717] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 874.090055] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating instance '4bf76ce0-b5d2-4184-888c-d0ef39878356' progress to 17 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 874.318339] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352615, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.351090] env[62914]: DEBUG nova.scheduler.client.report [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.465874] env[62914]: DEBUG oslo_concurrency.lockutils [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.466628] env[62914]: DEBUG oslo_concurrency.lockutils [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.466798] env[62914]: DEBUG oslo_concurrency.lockutils [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.467679] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44a5054-620c-43bd-977d-419ddbad4578 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.484878] env[62914]: DEBUG nova.virt.hardware [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.485123] env[62914]: DEBUG nova.virt.hardware [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.485289] env[62914]: DEBUG nova.virt.hardware [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.485476] env[62914]: DEBUG nova.virt.hardware [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.485624] env[62914]: DEBUG nova.virt.hardware [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.485770] env[62914]: DEBUG nova.virt.hardware [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.485979] env[62914]: DEBUG nova.virt.hardware [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.486158] env[62914]: DEBUG nova.virt.hardware [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.486327] env[62914]: DEBUG nova.virt.hardware [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.486491] env[62914]: DEBUG nova.virt.hardware [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.486662] env[62914]: DEBUG nova.virt.hardware [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.493097] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Reconfiguring VM to attach interface {{(pid=62914) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 874.493400] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4e56569-152c-4e6c-858a-bb13b71e6c28 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.512171] env[62914]: DEBUG oslo_vmware.api [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 874.512171] env[62914]: value = "task-1352617" [ 874.512171] env[62914]: _type = "Task" [ 874.512171] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.520632] env[62914]: DEBUG oslo_vmware.api [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352617, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.597288] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.597571] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.597979] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.598199] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.598398] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.598571] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.599274] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.599274] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.599274] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.599461] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.599673] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.605472] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5618097e-25b7-41ee-aeaa-5dd65123470c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.623064] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 874.623064] env[62914]: value = "task-1352618" [ 874.623064] env[62914]: _type = "Task" [ 874.623064] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.632148] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352618, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.818619] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352615, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.855858] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.329s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.856620] env[62914]: DEBUG nova.compute.manager [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 874.859580] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.406s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.861233] env[62914]: INFO nova.compute.claims [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 875.022831] env[62914]: DEBUG oslo_vmware.api [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352617, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.133380] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352618, 'name': ReconfigVM_Task, 'duration_secs': 0.224136} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.133686] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating instance '4bf76ce0-b5d2-4184-888c-d0ef39878356' progress to 33 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 875.320671] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352615, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.701361} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.320966] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 22de53a6-096b-4e8e-af91-f78588ea4c60/22de53a6-096b-4e8e-af91-f78588ea4c60.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 875.321259] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.321556] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef7360c1-0d99-4ed4-8695-121de0a1b909 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.328682] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 875.328682] env[62914]: value = "task-1352619" [ 875.328682] env[62914]: _type = "Task" [ 875.328682] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.336787] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352619, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.367739] env[62914]: DEBUG nova.compute.utils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 875.372982] env[62914]: DEBUG nova.compute.manager [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 875.373366] env[62914]: DEBUG nova.network.neutron [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 875.418356] env[62914]: DEBUG nova.policy [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '254bd9fe88344481abb4880c9052c362', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e682f282c61148abb7d06d2cfe624da9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 875.525377] env[62914]: DEBUG oslo_vmware.api [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352617, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.641029] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.641411] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.642437] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.642437] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.642437] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.642437] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.642754] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.643607] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.643607] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.643607] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.643800] env[62914]: DEBUG nova.virt.hardware [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.653441] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Reconfiguring VM instance instance-00000049 to detach disk 2000 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 875.654477] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-184fd694-7393-4d67-bd38-f7e8ec40c344 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.682345] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 875.682345] env[62914]: value = "task-1352620" [ 875.682345] env[62914]: _type = "Task" [ 875.682345] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.695126] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352620, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.740426] env[62914]: DEBUG nova.network.neutron [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Successfully created port: b705ac25-181c-430d-9661-805be91bb2a1 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 875.839207] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352619, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.114377} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.839584] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 875.840372] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8de6bc-c87f-4b55-b37e-acf83b5d3725 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.863421] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 22de53a6-096b-4e8e-af91-f78588ea4c60/22de53a6-096b-4e8e-af91-f78588ea4c60.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.863770] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da70c810-bbe8-48aa-8fa3-428407e47580 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.879131] env[62914]: DEBUG nova.compute.manager [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 875.890704] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 875.890704] env[62914]: value = "task-1352621" [ 875.890704] env[62914]: _type = "Task" [ 875.890704] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.900647] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352621, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.028308] env[62914]: DEBUG oslo_vmware.api [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352617, 'name': ReconfigVM_Task, 'duration_secs': 1.131111} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.028833] env[62914]: DEBUG oslo_concurrency.lockutils [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.029096] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Reconfigured VM to attach interface {{(pid=62914) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 876.087549] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714f0cec-eb30-48e3-8944-deef9f2fae92 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.096562] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1635aef-e25d-4877-bb2c-5db412939f00 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.130115] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1217fc6c-e5bd-44f9-a343-fa8ba4fcbe86 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.138118] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fcf294c-7016-4622-b5bd-09e62168802c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.153092] env[62914]: DEBUG nova.compute.provider_tree [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.193660] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352620, 'name': ReconfigVM_Task, 'duration_secs': 0.169341} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.193925] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Reconfigured VM instance instance-00000049 to detach disk 2000 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 876.194886] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772b8aa9-9b19-4a7f-9ec0-a78cb4545a1b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.218724] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 4bf76ce0-b5d2-4184-888c-d0ef39878356/4bf76ce0-b5d2-4184-888c-d0ef39878356.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.219058] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba048e52-2cb2-4f8e-b434-8624974a0db4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.239389] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 876.239389] env[62914]: value = "task-1352622" [ 876.239389] env[62914]: _type = "Task" [ 876.239389] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.249263] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352622, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.403215] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352621, 'name': ReconfigVM_Task, 'duration_secs': 0.302124} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.403534] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 22de53a6-096b-4e8e-af91-f78588ea4c60/22de53a6-096b-4e8e-af91-f78588ea4c60.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.404286] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e281d08c-553b-4d0f-bbc8-25fbf68c2edd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.412470] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 876.412470] env[62914]: value = "task-1352623" [ 876.412470] env[62914]: _type = "Task" [ 876.412470] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.421154] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352623, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.533908] env[62914]: DEBUG oslo_concurrency.lockutils [None req-70e4796b-52e1-4c6a-8ed4-d9f1ac98ef7c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.985s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.656585] env[62914]: DEBUG nova.scheduler.client.report [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.674638] env[62914]: DEBUG oslo_vmware.rw_handles [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527ad015-0af4-f987-187d-d7a213c98f1d/disk-0.vmdk. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 876.674890] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de44dcd-aefa-496b-a122-4a1a23d29ba7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.686441] env[62914]: DEBUG oslo_vmware.rw_handles [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527ad015-0af4-f987-187d-d7a213c98f1d/disk-0.vmdk is in state: ready. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 876.686701] env[62914]: ERROR oslo_vmware.rw_handles [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527ad015-0af4-f987-187d-d7a213c98f1d/disk-0.vmdk due to incomplete transfer. [ 876.687037] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a777fe4d-10f5-4c75-afc8-8919f83e2753 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.695779] env[62914]: DEBUG oslo_vmware.rw_handles [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527ad015-0af4-f987-187d-d7a213c98f1d/disk-0.vmdk. {{(pid=62914) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 876.695986] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Uploaded image 0b52b7b9-4343-488f-b3ec-34bf2b5cff88 to the Glance image server {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 876.698719] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Destroying the VM {{(pid=62914) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 876.698823] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-81c65f8d-dde1-4113-8118-9e0ed98f87a2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.704437] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 876.704437] env[62914]: value = "task-1352624" [ 876.704437] env[62914]: _type = "Task" [ 876.704437] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.713232] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352624, 'name': Destroy_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.749235] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352622, 'name': ReconfigVM_Task, 'duration_secs': 0.281227} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.749605] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 4bf76ce0-b5d2-4184-888c-d0ef39878356/4bf76ce0-b5d2-4184-888c-d0ef39878356.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.749806] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating instance '4bf76ce0-b5d2-4184-888c-d0ef39878356' progress to 50 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 876.893606] env[62914]: DEBUG nova.compute.manager [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 876.919086] env[62914]: DEBUG nova.virt.hardware [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.919376] env[62914]: DEBUG nova.virt.hardware [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.919537] env[62914]: DEBUG nova.virt.hardware [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.919725] env[62914]: DEBUG nova.virt.hardware [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.920156] env[62914]: DEBUG nova.virt.hardware [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.920156] env[62914]: DEBUG nova.virt.hardware [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.920318] env[62914]: DEBUG nova.virt.hardware [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.920481] env[62914]: DEBUG nova.virt.hardware [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.920648] env[62914]: DEBUG nova.virt.hardware [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.920812] env[62914]: DEBUG nova.virt.hardware [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.921166] env[62914]: DEBUG nova.virt.hardware [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.922016] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720bb567-1b73-480b-b24b-7f464156dd6b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.930087] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352623, 'name': Rename_Task, 'duration_secs': 0.157405} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.932148] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.932447] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c61456f1-e57b-424d-8a2d-10587b4cd7a4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.934881] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735df271-9ebd-4618-85d2-26740c211875 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.949637] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 876.949637] env[62914]: value = "task-1352625" [ 876.949637] env[62914]: _type = "Task" [ 876.949637] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.957733] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352625, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.162964] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.303s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.163570] env[62914]: DEBUG nova.compute.manager [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 877.167302] env[62914]: DEBUG oslo_concurrency.lockutils [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.444s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.167302] env[62914]: DEBUG nova.objects.instance [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Lazy-loading 'resources' on Instance uuid 9fab500c-344c-46cd-b060-c8f1aa0f1cba {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.215939] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352624, 'name': Destroy_Task} progress is 33%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.257434] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5011c1a3-cafc-4be5-98db-419d7acc95d3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.284445] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f188f5-f7b9-4a6a-b57b-2473a1ee6705 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.309235] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating instance '4bf76ce0-b5d2-4184-888c-d0ef39878356' progress to 67 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 877.461559] env[62914]: DEBUG nova.compute.manager [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Received event network-changed-bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.461841] env[62914]: DEBUG nova.compute.manager [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Refreshing instance network info cache due to event network-changed-bf4a3fa8-08a9-4890-abf6-fa02a60168a2. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 877.462205] env[62914]: DEBUG oslo_concurrency.lockutils [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] Acquiring lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.462280] env[62914]: DEBUG oslo_concurrency.lockutils [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] Acquired lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.462453] env[62914]: DEBUG nova.network.neutron [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Refreshing network info cache for port bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 877.468508] env[62914]: DEBUG oslo_vmware.api [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352625, 'name': PowerOnVM_Task, 'duration_secs': 0.506933} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.469013] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.469329] env[62914]: INFO nova.compute.manager [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Took 8.59 seconds to spawn the instance on the hypervisor. [ 877.469485] env[62914]: DEBUG nova.compute.manager [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.470353] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c63dc445-9a19-468c-8803-f019f3af6775 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.553381] env[62914]: DEBUG nova.network.neutron [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Successfully updated port: b705ac25-181c-430d-9661-805be91bb2a1 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.670201] env[62914]: DEBUG nova.compute.utils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 877.671766] env[62914]: DEBUG nova.compute.manager [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 877.671931] env[62914]: DEBUG nova.network.neutron [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 877.721423] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352624, 'name': Destroy_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.752787] env[62914]: DEBUG nova.policy [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '795b94761b144a1e99ed79838d9687ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9278818cebd842669b802a4e43e91774', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 877.845587] env[62914]: DEBUG nova.compute.manager [req-025a7c0b-982d-4cc5-8e25-682942b4430e req-e4db8b30-d9ca-4592-baea-b264f89214b8 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Received event network-changed-69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.845831] env[62914]: DEBUG nova.compute.manager [req-025a7c0b-982d-4cc5-8e25-682942b4430e req-e4db8b30-d9ca-4592-baea-b264f89214b8 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Refreshing instance network info cache due to event network-changed-69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 877.846951] env[62914]: DEBUG oslo_concurrency.lockutils [req-025a7c0b-982d-4cc5-8e25-682942b4430e req-e4db8b30-d9ca-4592-baea-b264f89214b8 service nova] Acquiring lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.847219] env[62914]: DEBUG oslo_concurrency.lockutils [req-025a7c0b-982d-4cc5-8e25-682942b4430e req-e4db8b30-d9ca-4592-baea-b264f89214b8 service nova] Acquired lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.848394] env[62914]: DEBUG nova.network.neutron [req-025a7c0b-982d-4cc5-8e25-682942b4430e req-e4db8b30-d9ca-4592-baea-b264f89214b8 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Refreshing network info cache for port 69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 877.862892] env[62914]: DEBUG nova.network.neutron [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Port 189c8622-547d-4da9-94da-564df3bc382d binding to destination host cpu-1 is already ACTIVE {{(pid=62914) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 877.966102] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c07211-0db3-43f4-8c6e-2e6d698d6c92 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.976602] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6bd8a9a-d719-43c9-939e-a7646d09b44e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.023688] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c515c847-dae8-408b-a0e9-ffe7df33d214 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.027106] env[62914]: INFO nova.compute.manager [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Took 22.10 seconds to build instance. [ 878.034477] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d97f248-3c21-466e-a143-c049981fe895 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.049227] env[62914]: DEBUG nova.compute.provider_tree [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.055987] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Acquiring lock "refresh_cache-47e6f142-bdb2-4397-9b94-d962e68ee6bb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.056169] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Acquired lock "refresh_cache-47e6f142-bdb2-4397-9b94-d962e68ee6bb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.056357] env[62914]: DEBUG nova.network.neutron [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 878.177959] env[62914]: DEBUG nova.compute.manager [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 878.223719] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352624, 'name': Destroy_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.233023] env[62914]: DEBUG nova.network.neutron [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Successfully created port: b5e84893-0b61-4ca8-b73b-1102b6477cf8 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 878.305017] env[62914]: DEBUG nova.network.neutron [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Updated VIF entry in instance network info cache for port bf4a3fa8-08a9-4890-abf6-fa02a60168a2. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 878.305017] env[62914]: DEBUG nova.network.neutron [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Updating instance_info_cache with network_info: [{"id": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "address": "fa:16:3e:c9:b5:2c", "network": {"id": "16e49b25-5bce-4cf2-8199-a865e6cd483e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1094957463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "28f67ce59c7f4151b37932bb890c4a5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf4a3fa8-08", "ovs_interfaceid": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.529656] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a3124dfd-5ac9-4532-b717-a2a6daee705e tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "22de53a6-096b-4e8e-af91-f78588ea4c60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.615s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.553375] env[62914]: DEBUG nova.scheduler.client.report [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 878.597275] env[62914]: DEBUG nova.network.neutron [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 878.720819] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352624, 'name': Destroy_Task, 'duration_secs': 1.749933} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.721111] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Destroyed the VM [ 878.721441] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Deleting Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 878.721702] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c094401f-365a-49b7-a539-66e6ecd5820a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.728584] env[62914]: DEBUG nova.network.neutron [req-025a7c0b-982d-4cc5-8e25-682942b4430e req-e4db8b30-d9ca-4592-baea-b264f89214b8 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Updated VIF entry in instance network info cache for port 69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 878.728998] env[62914]: DEBUG nova.network.neutron [req-025a7c0b-982d-4cc5-8e25-682942b4430e req-e4db8b30-d9ca-4592-baea-b264f89214b8 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Updating instance_info_cache with network_info: [{"id": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "address": "fa:16:3e:7a:f0:fa", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb12a55b6-c3", "ovs_interfaceid": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6", "address": "fa:16:3e:c1:ef:cc", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69f9fa9f-a0", "ovs_interfaceid": "69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.732373] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 878.732373] env[62914]: value = "task-1352626" [ 878.732373] env[62914]: _type = "Task" [ 878.732373] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.749301] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352626, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.793223] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.793491] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.793731] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.794280] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.794485] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.797295] env[62914]: INFO nova.compute.manager [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Terminating instance [ 878.799471] env[62914]: DEBUG nova.compute.manager [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 878.799568] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 878.800457] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8adec0c-1324-42c4-b176-14ab594703ad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.808259] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 878.808527] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1f1d68d-45e3-4e2b-ace4-10dd72ecbc3d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.811459] env[62914]: DEBUG nova.network.neutron [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Updating instance_info_cache with network_info: [{"id": "b705ac25-181c-430d-9661-805be91bb2a1", "address": "fa:16:3e:7f:8a:c1", "network": {"id": "6e94dba0-348f-4b8e-9009-318b74157c13", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-530503684-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e682f282c61148abb7d06d2cfe624da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb705ac25-18", "ovs_interfaceid": "b705ac25-181c-430d-9661-805be91bb2a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.812865] env[62914]: DEBUG oslo_concurrency.lockutils [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] Releasing lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.813116] env[62914]: DEBUG nova.compute.manager [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Received event network-changed-bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.813327] env[62914]: DEBUG nova.compute.manager [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Refreshing instance network info cache due to event network-changed-bf4a3fa8-08a9-4890-abf6-fa02a60168a2. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.813540] env[62914]: DEBUG oslo_concurrency.lockutils [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] Acquiring lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.813684] env[62914]: DEBUG oslo_concurrency.lockutils [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] Acquired lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.813846] env[62914]: DEBUG nova.network.neutron [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Refreshing network info cache for port bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 878.890829] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "4bf76ce0-b5d2-4184-888c-d0ef39878356-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.891226] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "4bf76ce0-b5d2-4184-888c-d0ef39878356-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.891529] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "4bf76ce0-b5d2-4184-888c-d0ef39878356-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.061189] env[62914]: DEBUG oslo_concurrency.lockutils [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.893s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.065781] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.576s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.065781] env[62914]: DEBUG nova.objects.instance [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Lazy-loading 'resources' on Instance uuid f80ee46d-9209-4b60-9e5a-f7c8764f4ccb {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.097470] env[62914]: INFO nova.scheduler.client.report [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Deleted allocations for instance 9fab500c-344c-46cd-b060-c8f1aa0f1cba [ 879.151830] env[62914]: DEBUG oslo_concurrency.lockutils [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "interface-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.151830] env[62914]: DEBUG oslo_concurrency.lockutils [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.190426] env[62914]: DEBUG nova.compute.manager [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 879.221352] env[62914]: DEBUG nova.virt.hardware [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.221636] env[62914]: DEBUG nova.virt.hardware [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.221797] env[62914]: DEBUG nova.virt.hardware [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.221977] env[62914]: DEBUG nova.virt.hardware [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.222138] env[62914]: DEBUG nova.virt.hardware [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.222275] env[62914]: DEBUG nova.virt.hardware [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.222475] env[62914]: DEBUG nova.virt.hardware [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.222629] env[62914]: DEBUG nova.virt.hardware [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.222796] env[62914]: DEBUG nova.virt.hardware [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.222957] env[62914]: DEBUG nova.virt.hardware [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.223182] env[62914]: DEBUG nova.virt.hardware [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.224121] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fdd6174-311a-4fce-b03e-56bf5985789e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.233293] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef00c51-7197-4b74-876d-df0aec5cf0d9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.237250] env[62914]: DEBUG oslo_concurrency.lockutils [req-025a7c0b-982d-4cc5-8e25-682942b4430e req-e4db8b30-d9ca-4592-baea-b264f89214b8 service nova] Releasing lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.244146] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 879.244418] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 879.244627] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleting the datastore file [datastore1] 43ee8948-8805-4d48-bd45-e93d2e2eb05d {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 879.258873] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-61b53979-b1f6-45c2-938e-9f4bada09f9a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.261120] env[62914]: DEBUG oslo_vmware.api [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352626, 'name': RemoveSnapshot_Task, 'duration_secs': 0.43108} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.261403] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Deleted Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 879.261657] env[62914]: INFO nova.compute.manager [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Took 17.90 seconds to snapshot the instance on the hypervisor. [ 879.269538] env[62914]: DEBUG oslo_vmware.api [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 879.269538] env[62914]: value = "task-1352628" [ 879.269538] env[62914]: _type = "Task" [ 879.269538] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.278324] env[62914]: DEBUG oslo_vmware.api [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.316392] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Releasing lock "refresh_cache-47e6f142-bdb2-4397-9b94-d962e68ee6bb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.316820] env[62914]: DEBUG nova.compute.manager [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Instance network_info: |[{"id": "b705ac25-181c-430d-9661-805be91bb2a1", "address": "fa:16:3e:7f:8a:c1", "network": {"id": "6e94dba0-348f-4b8e-9009-318b74157c13", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-530503684-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e682f282c61148abb7d06d2cfe624da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb705ac25-18", "ovs_interfaceid": "b705ac25-181c-430d-9661-805be91bb2a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 879.319215] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:8a:c1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1f996252-e329-42bd-a897-446dfe2b81cd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b705ac25-181c-430d-9661-805be91bb2a1', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 879.329040] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Creating folder: Project (e682f282c61148abb7d06d2cfe624da9). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 879.329040] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-588f6a4a-6919-4ac6-acbf-4f49f7e5cc37 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.340323] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Created folder: Project (e682f282c61148abb7d06d2cfe624da9) in parent group-v288131. [ 879.340544] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Creating folder: Instances. Parent ref: group-v288214. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 879.340804] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc9b3d32-0947-4c5b-b1a2-9dbe968eb935 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.351635] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Created folder: Instances in parent group-v288214. [ 879.351635] env[62914]: DEBUG oslo.service.loopingcall [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 879.351790] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 879.352817] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-053cfb44-6e44-4692-a221-a599cd5af779 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.379973] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 879.379973] env[62914]: value = "task-1352631" [ 879.379973] env[62914]: _type = "Task" [ 879.379973] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.390049] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352631, 'name': CreateVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.498452] env[62914]: DEBUG nova.compute.manager [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Received event network-vif-plugged-b705ac25-181c-430d-9661-805be91bb2a1 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.498767] env[62914]: DEBUG oslo_concurrency.lockutils [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] Acquiring lock "47e6f142-bdb2-4397-9b94-d962e68ee6bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.499070] env[62914]: DEBUG oslo_concurrency.lockutils [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] Lock "47e6f142-bdb2-4397-9b94-d962e68ee6bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.499319] env[62914]: DEBUG oslo_concurrency.lockutils [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] Lock "47e6f142-bdb2-4397-9b94-d962e68ee6bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.499576] env[62914]: DEBUG nova.compute.manager [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] No waiting events found dispatching network-vif-plugged-b705ac25-181c-430d-9661-805be91bb2a1 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 879.499820] env[62914]: WARNING nova.compute.manager [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Received unexpected event network-vif-plugged-b705ac25-181c-430d-9661-805be91bb2a1 for instance with vm_state building and task_state spawning. [ 879.500114] env[62914]: DEBUG nova.compute.manager [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Received event network-changed-b705ac25-181c-430d-9661-805be91bb2a1 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.500355] env[62914]: DEBUG nova.compute.manager [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Refreshing instance network info cache due to event network-changed-b705ac25-181c-430d-9661-805be91bb2a1. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 879.500659] env[62914]: DEBUG oslo_concurrency.lockutils [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] Acquiring lock "refresh_cache-47e6f142-bdb2-4397-9b94-d962e68ee6bb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.500818] env[62914]: DEBUG oslo_concurrency.lockutils [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] Acquired lock "refresh_cache-47e6f142-bdb2-4397-9b94-d962e68ee6bb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.501055] env[62914]: DEBUG nova.network.neutron [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Refreshing network info cache for port b705ac25-181c-430d-9661-805be91bb2a1 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.596304] env[62914]: DEBUG nova.network.neutron [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Updated VIF entry in instance network info cache for port bf4a3fa8-08a9-4890-abf6-fa02a60168a2. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 879.596304] env[62914]: DEBUG nova.network.neutron [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Updating instance_info_cache with network_info: [{"id": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "address": "fa:16:3e:c9:b5:2c", "network": {"id": "16e49b25-5bce-4cf2-8199-a865e6cd483e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1094957463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "28f67ce59c7f4151b37932bb890c4a5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf4a3fa8-08", "ovs_interfaceid": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.607555] env[62914]: DEBUG oslo_concurrency.lockutils [None req-58e14cf8-6192-44a3-9f82-e59404af56aa tempest-ServersTestManualDisk-298714815 tempest-ServersTestManualDisk-298714815-project-member] Lock "9fab500c-344c-46cd-b060-c8f1aa0f1cba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.461s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.659155] env[62914]: DEBUG oslo_concurrency.lockutils [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.659155] env[62914]: DEBUG oslo_concurrency.lockutils [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.666234] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c5990a-f528-43b4-b981-60d3c6bf577a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.669552] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "aec39e18-7796-4be3-af74-478df3a78f8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.669943] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "aec39e18-7796-4be3-af74-478df3a78f8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.693791] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86dbe2b2-aa5e-41a0-9dc4-aa7df79900cd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.726944] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Reconfiguring VM to detach interface {{(pid=62914) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 879.730095] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64eb40e4-d0e0-420b-a226-8f88451f310c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.757864] env[62914]: DEBUG oslo_vmware.api [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 879.757864] env[62914]: value = "task-1352632" [ 879.757864] env[62914]: _type = "Task" [ 879.757864] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.766213] env[62914]: DEBUG nova.compute.manager [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Instance disappeared during snapshot {{(pid=62914) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 879.772013] env[62914]: DEBUG oslo_vmware.api [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352632, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.782393] env[62914]: DEBUG oslo_vmware.api [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352628, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.314702} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.782393] env[62914]: DEBUG nova.compute.manager [None req-4d8f29fa-1745-4d7d-864e-af5103fa9bd7 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Image not found during clean up 0b52b7b9-4343-488f-b3ec-34bf2b5cff88 {{(pid=62914) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 879.785397] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 879.785577] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 879.785795] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 879.785937] env[62914]: INFO nova.compute.manager [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Took 0.99 seconds to destroy the instance on the hypervisor. [ 879.786211] env[62914]: DEBUG oslo.service.loopingcall [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 879.786808] env[62914]: DEBUG nova.compute.manager [-] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 879.786933] env[62914]: DEBUG nova.network.neutron [-] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 879.855956] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9069c61-b5b4-43b4-8509-25b0eab56c0e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.864336] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e444ae71-bc0c-4228-b776-7e2cd39ff1cc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.908367] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e6bef8-0ec8-406a-946f-dcfe6a9665a8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.914290] env[62914]: DEBUG nova.compute.manager [req-0ba271c3-ae9a-428e-9ccd-afe471245b09 req-3650a331-5a32-4ba8-b706-946282432d68 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Received event network-changed-bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.914431] env[62914]: DEBUG nova.compute.manager [req-0ba271c3-ae9a-428e-9ccd-afe471245b09 req-3650a331-5a32-4ba8-b706-946282432d68 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Refreshing instance network info cache due to event network-changed-bf4a3fa8-08a9-4890-abf6-fa02a60168a2. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 879.914589] env[62914]: DEBUG oslo_concurrency.lockutils [req-0ba271c3-ae9a-428e-9ccd-afe471245b09 req-3650a331-5a32-4ba8-b706-946282432d68 service nova] Acquiring lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.921333] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352631, 'name': CreateVM_Task, 'duration_secs': 0.449041} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.923409] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 879.924201] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.924371] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.924839] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 879.925989] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-725f33b6-6a4d-4432-8087-cff128b0bc37 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.929953] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13efd83f-1665-4a05-baeb-f0711c8fbe47 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.937907] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Waiting for the task: (returnval){ [ 879.937907] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52eab563-3fc1-4f7c-e309-920e187625ff" [ 879.937907] env[62914]: _type = "Task" [ 879.937907] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.949387] env[62914]: DEBUG nova.compute.provider_tree [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.962550] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52eab563-3fc1-4f7c-e309-920e187625ff, 'name': SearchDatastore_Task, 'duration_secs': 0.014855} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.966021] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.966021] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.966021] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.966021] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.966021] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.966021] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-643670d1-ce15-44e8-9edb-8fd52df776f0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.973985] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.974222] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 879.974984] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8b15c3c-5071-4c22-b341-9f750bcbc9b2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.981241] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Waiting for the task: (returnval){ [ 879.981241] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d9dd12-906c-5c01-0d5a-e1a7c9dca3a9" [ 879.981241] env[62914]: _type = "Task" [ 879.981241] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.989063] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d9dd12-906c-5c01-0d5a-e1a7c9dca3a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.993865] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.994097] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.994322] env[62914]: DEBUG nova.network.neutron [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 880.091381] env[62914]: DEBUG nova.network.neutron [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Successfully updated port: b5e84893-0b61-4ca8-b73b-1102b6477cf8 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 880.096452] env[62914]: DEBUG oslo_concurrency.lockutils [req-e8e23591-12cf-4c96-91d9-0f6b447854a7 req-6240e5fa-16cd-4863-b08f-2aafd5ba5952 service nova] Releasing lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.096878] env[62914]: DEBUG oslo_concurrency.lockutils [req-0ba271c3-ae9a-428e-9ccd-afe471245b09 req-3650a331-5a32-4ba8-b706-946282432d68 service nova] Acquired lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.097084] env[62914]: DEBUG nova.network.neutron [req-0ba271c3-ae9a-428e-9ccd-afe471245b09 req-3650a331-5a32-4ba8-b706-946282432d68 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Refreshing network info cache for port bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 880.176021] env[62914]: DEBUG nova.compute.manager [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 880.273306] env[62914]: DEBUG oslo_vmware.api [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352632, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.459367] env[62914]: DEBUG nova.scheduler.client.report [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 880.494222] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d9dd12-906c-5c01-0d5a-e1a7c9dca3a9, 'name': SearchDatastore_Task, 'duration_secs': 0.013124} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.495312] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50fe9b45-18c9-46ae-beb7-bd6f7b4c7aee {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.501805] env[62914]: DEBUG oslo_vmware.rw_handles [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5298fab9-1ad3-f8fe-f920-0b2ae67e3957/disk-0.vmdk. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 880.504459] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591fefc2-bd21-4488-8ca8-537f2670fd62 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.510534] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Waiting for the task: (returnval){ [ 880.510534] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]525cb716-51ee-7a10-1464-6af8346d6add" [ 880.510534] env[62914]: _type = "Task" [ 880.510534] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.512613] env[62914]: DEBUG oslo_vmware.rw_handles [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5298fab9-1ad3-f8fe-f920-0b2ae67e3957/disk-0.vmdk is in state: ready. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 880.516023] env[62914]: ERROR oslo_vmware.rw_handles [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5298fab9-1ad3-f8fe-f920-0b2ae67e3957/disk-0.vmdk due to incomplete transfer. [ 880.516401] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9c3a8107-3ab1-45d4-8ce2-9f04f6d95267 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.523319] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]525cb716-51ee-7a10-1464-6af8346d6add, 'name': SearchDatastore_Task} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.524141] env[62914]: DEBUG nova.network.neutron [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Updated VIF entry in instance network info cache for port b705ac25-181c-430d-9661-805be91bb2a1. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 880.524478] env[62914]: DEBUG nova.network.neutron [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Updating instance_info_cache with network_info: [{"id": "b705ac25-181c-430d-9661-805be91bb2a1", "address": "fa:16:3e:7f:8a:c1", "network": {"id": "6e94dba0-348f-4b8e-9009-318b74157c13", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-530503684-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e682f282c61148abb7d06d2cfe624da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb705ac25-18", "ovs_interfaceid": "b705ac25-181c-430d-9661-805be91bb2a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.529017] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.529017] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 47e6f142-bdb2-4397-9b94-d962e68ee6bb/47e6f142-bdb2-4397-9b94-d962e68ee6bb.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 880.529017] env[62914]: DEBUG oslo_vmware.rw_handles [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5298fab9-1ad3-f8fe-f920-0b2ae67e3957/disk-0.vmdk. {{(pid=62914) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 880.529017] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Uploaded image 8e275538-804d-43f7-8258-35fc043e5a94 to the Glance image server {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 880.529017] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Destroying the VM {{(pid=62914) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 880.529992] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5cf5a128-b1f9-4931-afe5-c7fe2a33326c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.531292] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-198549a4-e08a-41be-8f3e-e8815a9b8d1e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.538166] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Waiting for the task: (returnval){ [ 880.538166] env[62914]: value = "task-1352633" [ 880.538166] env[62914]: _type = "Task" [ 880.538166] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.539445] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 880.539445] env[62914]: value = "task-1352634" [ 880.539445] env[62914]: _type = "Task" [ 880.539445] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.552302] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352633, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.555583] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352634, 'name': Destroy_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.570937] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "22de53a6-096b-4e8e-af91-f78588ea4c60" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.571297] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "22de53a6-096b-4e8e-af91-f78588ea4c60" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.594177] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "refresh_cache-c9b94435-6546-464a-b3ba-c9e685157059" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.594352] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "refresh_cache-c9b94435-6546-464a-b3ba-c9e685157059" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.594476] env[62914]: DEBUG nova.network.neutron [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 880.624375] env[62914]: DEBUG nova.network.neutron [-] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.703172] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.771045] env[62914]: DEBUG oslo_vmware.api [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352632, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.839393] env[62914]: DEBUG nova.network.neutron [req-0ba271c3-ae9a-428e-9ccd-afe471245b09 req-3650a331-5a32-4ba8-b706-946282432d68 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Updated VIF entry in instance network info cache for port bf4a3fa8-08a9-4890-abf6-fa02a60168a2. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 880.839837] env[62914]: DEBUG nova.network.neutron [req-0ba271c3-ae9a-428e-9ccd-afe471245b09 req-3650a331-5a32-4ba8-b706-946282432d68 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Updating instance_info_cache with network_info: [{"id": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "address": "fa:16:3e:c9:b5:2c", "network": {"id": "16e49b25-5bce-4cf2-8199-a865e6cd483e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1094957463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "28f67ce59c7f4151b37932bb890c4a5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf4a3fa8-08", "ovs_interfaceid": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.964378] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.901s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.966749] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.264s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.969063] env[62914]: INFO nova.compute.claims [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.988633] env[62914]: INFO nova.scheduler.client.report [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Deleted allocations for instance f80ee46d-9209-4b60-9e5a-f7c8764f4ccb [ 881.030499] env[62914]: DEBUG oslo_concurrency.lockutils [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] Releasing lock "refresh_cache-47e6f142-bdb2-4397-9b94-d962e68ee6bb" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.030756] env[62914]: DEBUG nova.compute.manager [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Received event network-changed-bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.030932] env[62914]: DEBUG nova.compute.manager [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Refreshing instance network info cache due to event network-changed-bf4a3fa8-08a9-4890-abf6-fa02a60168a2. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 881.031170] env[62914]: DEBUG oslo_concurrency.lockutils [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] Acquiring lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.052013] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352633, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.054914] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352634, 'name': Destroy_Task} progress is 33%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.075642] env[62914]: DEBUG nova.compute.utils [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 881.126956] env[62914]: INFO nova.compute.manager [-] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Took 1.34 seconds to deallocate network for instance. [ 881.129359] env[62914]: DEBUG nova.network.neutron [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 881.162268] env[62914]: DEBUG nova.network.neutron [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating instance_info_cache with network_info: [{"id": "189c8622-547d-4da9-94da-564df3bc382d", "address": "fa:16:3e:b6:1a:f0", "network": {"id": "21b6ff73-feab-478c-848a-313ea0acbfc2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1273337660-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "619be7b3bf1445b68ba7adefc98f8782", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap189c8622-54", "ovs_interfaceid": "189c8622-547d-4da9-94da-564df3bc382d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.269540] env[62914]: DEBUG oslo_vmware.api [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352632, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.274817] env[62914]: DEBUG nova.network.neutron [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Updating instance_info_cache with network_info: [{"id": "b5e84893-0b61-4ca8-b73b-1102b6477cf8", "address": "fa:16:3e:fe:5b:19", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5e84893-0b", "ovs_interfaceid": "b5e84893-0b61-4ca8-b73b-1102b6477cf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.342710] env[62914]: DEBUG oslo_concurrency.lockutils [req-0ba271c3-ae9a-428e-9ccd-afe471245b09 req-3650a331-5a32-4ba8-b706-946282432d68 service nova] Releasing lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.343174] env[62914]: DEBUG oslo_concurrency.lockutils [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] Acquired lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.343360] env[62914]: DEBUG nova.network.neutron [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Refreshing network info cache for port bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 881.498092] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6ffb4723-6198-461b-bb8e-0b0c87343734 tempest-ServerGroupTestJSON-1444352863 tempest-ServerGroupTestJSON-1444352863-project-member] Lock "f80ee46d-9209-4b60-9e5a-f7c8764f4ccb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.953s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.552522] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352633, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526873} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.555401] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 47e6f142-bdb2-4397-9b94-d962e68ee6bb/47e6f142-bdb2-4397-9b94-d962e68ee6bb.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 881.555633] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.555927] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352634, 'name': Destroy_Task, 'duration_secs': 0.589633} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.556112] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-24e5b21b-8f08-4ca4-9039-695034c25fbe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.557864] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Destroyed the VM [ 881.558147] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Deleting Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 881.558387] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cdf46c2d-4032-4514-8be9-afb5846f6869 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.563778] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Waiting for the task: (returnval){ [ 881.563778] env[62914]: value = "task-1352635" [ 881.563778] env[62914]: _type = "Task" [ 881.563778] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.564947] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 881.564947] env[62914]: value = "task-1352636" [ 881.564947] env[62914]: _type = "Task" [ 881.564947] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.575193] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352636, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.578745] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "22de53a6-096b-4e8e-af91-f78588ea4c60" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.579312] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352635, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.639977] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.665097] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquiring lock "29680cfb-8df3-4ff7-b6ee-0982577cc9cf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.665373] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Lock "29680cfb-8df3-4ff7-b6ee-0982577cc9cf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.665698] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquiring lock "29680cfb-8df3-4ff7-b6ee-0982577cc9cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.665784] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Lock "29680cfb-8df3-4ff7-b6ee-0982577cc9cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.665950] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Lock "29680cfb-8df3-4ff7-b6ee-0982577cc9cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.667845] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.671783] env[62914]: INFO nova.compute.manager [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Terminating instance [ 881.677586] env[62914]: DEBUG nova.compute.manager [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 881.677786] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 881.678753] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4423cefe-b7ba-4935-bf11-1cd7a0675a7d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.688561] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 881.688938] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f20110d-f892-4078-b331-95281ec6cd3b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.694814] env[62914]: DEBUG oslo_vmware.api [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 881.694814] env[62914]: value = "task-1352637" [ 881.694814] env[62914]: _type = "Task" [ 881.694814] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.702779] env[62914]: DEBUG oslo_vmware.api [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352637, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.742392] env[62914]: DEBUG nova.compute.manager [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Received event network-vif-plugged-b5e84893-0b61-4ca8-b73b-1102b6477cf8 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.742621] env[62914]: DEBUG oslo_concurrency.lockutils [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] Acquiring lock "c9b94435-6546-464a-b3ba-c9e685157059-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.742828] env[62914]: DEBUG oslo_concurrency.lockutils [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] Lock "c9b94435-6546-464a-b3ba-c9e685157059-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.743008] env[62914]: DEBUG oslo_concurrency.lockutils [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] Lock "c9b94435-6546-464a-b3ba-c9e685157059-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.744016] env[62914]: DEBUG nova.compute.manager [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] [instance: c9b94435-6546-464a-b3ba-c9e685157059] No waiting events found dispatching network-vif-plugged-b5e84893-0b61-4ca8-b73b-1102b6477cf8 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 881.744016] env[62914]: WARNING nova.compute.manager [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Received unexpected event network-vif-plugged-b5e84893-0b61-4ca8-b73b-1102b6477cf8 for instance with vm_state building and task_state spawning. [ 881.744016] env[62914]: DEBUG nova.compute.manager [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Received event network-changed-b5e84893-0b61-4ca8-b73b-1102b6477cf8 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.744347] env[62914]: DEBUG nova.compute.manager [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Refreshing instance network info cache due to event network-changed-b5e84893-0b61-4ca8-b73b-1102b6477cf8. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 881.744513] env[62914]: DEBUG oslo_concurrency.lockutils [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] Acquiring lock "refresh_cache-c9b94435-6546-464a-b3ba-c9e685157059" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.769810] env[62914]: DEBUG oslo_vmware.api [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352632, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.777860] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "refresh_cache-c9b94435-6546-464a-b3ba-c9e685157059" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.778184] env[62914]: DEBUG nova.compute.manager [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Instance network_info: |[{"id": "b5e84893-0b61-4ca8-b73b-1102b6477cf8", "address": "fa:16:3e:fe:5b:19", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5e84893-0b", "ovs_interfaceid": "b5e84893-0b61-4ca8-b73b-1102b6477cf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 881.778487] env[62914]: DEBUG oslo_concurrency.lockutils [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] Acquired lock "refresh_cache-c9b94435-6546-464a-b3ba-c9e685157059" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.778673] env[62914]: DEBUG nova.network.neutron [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Refreshing network info cache for port b5e84893-0b61-4ca8-b73b-1102b6477cf8 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 881.779925] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:5b:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '32faf59b-014c-4f1f-8331-40df95bf741f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b5e84893-0b61-4ca8-b73b-1102b6477cf8', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.792810] env[62914]: DEBUG oslo.service.loopingcall [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.796077] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 881.796908] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f614373e-150b-4fd7-bf34-70c4ef0781dc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.817700] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.817700] env[62914]: value = "task-1352638" [ 881.817700] env[62914]: _type = "Task" [ 881.817700] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.827155] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352638, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.087836] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352635, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069565} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.088271] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.089272] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a9aab7-0caa-4505-a2f6-ed25842af357 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.097088] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352636, 'name': RemoveSnapshot_Task} progress is 15%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.121711] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 47e6f142-bdb2-4397-9b94-d962e68ee6bb/47e6f142-bdb2-4397-9b94-d962e68ee6bb.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.131017] env[62914]: DEBUG nova.network.neutron [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Updated VIF entry in instance network info cache for port bf4a3fa8-08a9-4890-abf6-fa02a60168a2. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 882.131017] env[62914]: DEBUG nova.network.neutron [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Updating instance_info_cache with network_info: [{"id": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "address": "fa:16:3e:c9:b5:2c", "network": {"id": "16e49b25-5bce-4cf2-8199-a865e6cd483e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1094957463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "28f67ce59c7f4151b37932bb890c4a5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9f208df-1fb5-4403-9796-7fd19e4bfb85", "external-id": "cl2-zone-400", "segmentation_id": 400, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf4a3fa8-08", "ovs_interfaceid": "bf4a3fa8-08a9-4890-abf6-fa02a60168a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.133172] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9142dd9-81ab-4c3a-ac60-0ea5675f87d2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.150032] env[62914]: DEBUG oslo_concurrency.lockutils [req-3d7e8cef-ba6c-4e06-aec4-4c76577ec3ea req-22a7f575-f793-4523-ad16-a89826fdc4fd service nova] Releasing lock "refresh_cache-29680cfb-8df3-4ff7-b6ee-0982577cc9cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.157044] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Waiting for the task: (returnval){ [ 882.157044] env[62914]: value = "task-1352639" [ 882.157044] env[62914]: _type = "Task" [ 882.157044] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.171146] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352639, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.207429] env[62914]: DEBUG oslo_vmware.api [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352637, 'name': PowerOffVM_Task, 'duration_secs': 0.337303} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.211331] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 882.211590] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 882.213547] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f18f3709-08fe-44ea-be25-fd13a580c3cc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.216407] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10762b84-a84f-4f2e-9a1f-a34b88e9036a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.245782] env[62914]: DEBUG nova.network.neutron [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Updated VIF entry in instance network info cache for port b5e84893-0b61-4ca8-b73b-1102b6477cf8. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 882.246227] env[62914]: DEBUG nova.network.neutron [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Updating instance_info_cache with network_info: [{"id": "b5e84893-0b61-4ca8-b73b-1102b6477cf8", "address": "fa:16:3e:fe:5b:19", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5e84893-0b", "ovs_interfaceid": "b5e84893-0b61-4ca8-b73b-1102b6477cf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.248222] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f99c9a-954b-4702-b9c7-8218a0942761 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.257880] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating instance '4bf76ce0-b5d2-4184-888c-d0ef39878356' progress to 83 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 882.268425] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238f211d-6652-4618-8ee9-27950ccd8b7d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.274943] env[62914]: DEBUG oslo_vmware.api [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352632, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.280378] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10fc268e-9119-48f3-a18f-2e289ea29e19 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.320682] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473df152-43be-408f-8e24-595dc2fb65fa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.327027] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 882.327027] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 882.327027] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Deleting the datastore file [datastore2] 29680cfb-8df3-4ff7-b6ee-0982577cc9cf {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.327829] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0cdbb032-b2c1-43d4-9bb6-f496a17cdd39 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.341065] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-075b940e-73a3-4561-932a-c75244696ab0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.346303] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352638, 'name': CreateVM_Task, 'duration_secs': 0.485027} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.346836] env[62914]: DEBUG oslo_vmware.api [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for the task: (returnval){ [ 882.346836] env[62914]: value = "task-1352641" [ 882.346836] env[62914]: _type = "Task" [ 882.346836] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.347352] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 882.348667] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.348917] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.349406] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.365301] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e359074-6185-4ef6-a092-3b24f6236662 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.368520] env[62914]: DEBUG nova.compute.provider_tree [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.376138] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 882.376138] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]525d6fe1-ba70-a2dc-2b91-8be037628d3b" [ 882.376138] env[62914]: _type = "Task" [ 882.376138] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.377144] env[62914]: DEBUG oslo_vmware.api [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352641, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.387877] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]525d6fe1-ba70-a2dc-2b91-8be037628d3b, 'name': SearchDatastore_Task, 'duration_secs': 0.010467} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.388285] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.388582] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.389541] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.389541] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.389541] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.389541] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e4aaa0e4-282c-4aa6-b8b6-773c2b374edc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.399135] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.399135] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 882.399135] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0792fffb-cc64-406f-b45d-974c1c87d52a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.405117] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 882.405117] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ca70a4-81b9-be65-3e18-1c821f29208b" [ 882.405117] env[62914]: _type = "Task" [ 882.405117] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.412452] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ca70a4-81b9-be65-3e18-1c821f29208b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.585661] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352636, 'name': RemoveSnapshot_Task} progress is 70%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.670988] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352639, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.689013] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "22de53a6-096b-4e8e-af91-f78588ea4c60" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.689292] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "22de53a6-096b-4e8e-af91-f78588ea4c60" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.689559] env[62914]: INFO nova.compute.manager [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Attaching volume 3b0294c0-2d37-450a-a517-8c71f44b449e to /dev/sdb [ 882.748120] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d1049b8-54ee-4ce7-8e77-06bc0e24d912 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.756491] env[62914]: DEBUG oslo_concurrency.lockutils [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] Releasing lock "refresh_cache-c9b94435-6546-464a-b3ba-c9e685157059" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.756491] env[62914]: DEBUG nova.compute.manager [req-d13b02a8-4e12-40df-8751-425212439a2a req-120b935b-cb37-42ed-a62c-4279a646a5de service nova] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Received event network-vif-deleted-cf125424-b194-49ac-886d-a8b053ae7a77 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.757496] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a67482-6af4-405d-a7f2-a8571dead986 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.766724] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 882.769387] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5259ee84-48b7-4505-82c4-e8cd4ec34e94 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.776816] env[62914]: DEBUG nova.virt.block_device [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Updating existing volume attachment record: 4a113649-1c60-4e0f-91d1-55a3398634c3 {{(pid=62914) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 882.785180] env[62914]: DEBUG oslo_vmware.api [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352632, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.786447] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 882.786447] env[62914]: value = "task-1352642" [ 882.786447] env[62914]: _type = "Task" [ 882.786447] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.797180] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352642, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.861821] env[62914]: DEBUG oslo_vmware.api [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Task: {'id': task-1352641, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.33234} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.861821] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.861821] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 882.861821] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 882.862015] env[62914]: INFO nova.compute.manager [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Took 1.18 seconds to destroy the instance on the hypervisor. [ 882.863062] env[62914]: DEBUG oslo.service.loopingcall [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.863062] env[62914]: DEBUG nova.compute.manager [-] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 882.863062] env[62914]: DEBUG nova.network.neutron [-] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 882.876171] env[62914]: DEBUG nova.scheduler.client.report [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.915572] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ca70a4-81b9-be65-3e18-1c821f29208b, 'name': SearchDatastore_Task, 'duration_secs': 0.01001} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.916640] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66019e44-ae6d-4f59-bf7c-b4cf730fa397 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.923754] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 882.923754] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5221d89b-00a4-9815-d8b9-fa82cc0eff72" [ 882.923754] env[62914]: _type = "Task" [ 882.923754] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.933454] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5221d89b-00a4-9815-d8b9-fa82cc0eff72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.082197] env[62914]: DEBUG oslo_vmware.api [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352636, 'name': RemoveSnapshot_Task, 'duration_secs': 1.404796} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.082554] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Deleted Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 883.082799] env[62914]: INFO nova.compute.manager [None req-1bd2813e-c9e5-4406-813e-08c93aff8342 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Took 15.33 seconds to snapshot the instance on the hypervisor. [ 883.173700] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352639, 'name': ReconfigVM_Task, 'duration_secs': 0.526388} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.173982] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 47e6f142-bdb2-4397-9b94-d962e68ee6bb/47e6f142-bdb2-4397-9b94-d962e68ee6bb.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.175129] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e6fa849-a05c-4266-bbac-688579422a4d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.179365] env[62914]: DEBUG nova.compute.manager [req-6def7d4f-4093-49b9-ad44-5ad3ae623542 req-50ffb618-6406-4902-bc8c-a09e15d623e0 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Received event network-vif-deleted-bf4a3fa8-08a9-4890-abf6-fa02a60168a2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.179557] env[62914]: INFO nova.compute.manager [req-6def7d4f-4093-49b9-ad44-5ad3ae623542 req-50ffb618-6406-4902-bc8c-a09e15d623e0 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Neutron deleted interface bf4a3fa8-08a9-4890-abf6-fa02a60168a2; detaching it from the instance and deleting it from the info cache [ 883.179738] env[62914]: DEBUG nova.network.neutron [req-6def7d4f-4093-49b9-ad44-5ad3ae623542 req-50ffb618-6406-4902-bc8c-a09e15d623e0 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.186369] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Waiting for the task: (returnval){ [ 883.186369] env[62914]: value = "task-1352646" [ 883.186369] env[62914]: _type = "Task" [ 883.186369] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.196544] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352646, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.273497] env[62914]: DEBUG oslo_vmware.api [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352632, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.298333] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352642, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.380544] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.414s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.381100] env[62914]: DEBUG nova.compute.manager [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 883.383947] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.744s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.384199] env[62914]: DEBUG nova.objects.instance [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lazy-loading 'resources' on Instance uuid 43ee8948-8805-4d48-bd45-e93d2e2eb05d {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.434883] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5221d89b-00a4-9815-d8b9-fa82cc0eff72, 'name': SearchDatastore_Task, 'duration_secs': 0.019057} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.435271] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.435559] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] c9b94435-6546-464a-b3ba-c9e685157059/c9b94435-6546-464a-b3ba-c9e685157059.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 883.435822] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f854772-fdbc-4b1d-a193-c32b5e805047 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.442555] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 883.442555] env[62914]: value = "task-1352647" [ 883.442555] env[62914]: _type = "Task" [ 883.442555] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.452089] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352647, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.655287] env[62914]: DEBUG nova.network.neutron [-] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.682256] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7799fc94-4d8d-48c0-abbe-76b71a1806f5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.695651] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8be8dc4-6593-4248-9300-f9b310525b6c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.712462] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352646, 'name': Rename_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.730521] env[62914]: DEBUG nova.compute.manager [req-6def7d4f-4093-49b9-ad44-5ad3ae623542 req-50ffb618-6406-4902-bc8c-a09e15d623e0 service nova] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Detach interface failed, port_id=bf4a3fa8-08a9-4890-abf6-fa02a60168a2, reason: Instance 29680cfb-8df3-4ff7-b6ee-0982577cc9cf could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 883.775377] env[62914]: DEBUG oslo_vmware.api [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352632, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.804031] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352642, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.888536] env[62914]: DEBUG nova.compute.utils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 883.895206] env[62914]: DEBUG nova.compute.manager [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 883.895206] env[62914]: DEBUG nova.network.neutron [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 883.956517] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352647, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488828} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.959231] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] c9b94435-6546-464a-b3ba-c9e685157059/c9b94435-6546-464a-b3ba-c9e685157059.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 883.959465] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 883.959912] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc7dc713-f470-461b-be0a-424945411178 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.963208] env[62914]: DEBUG nova.policy [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41c19b40819c49d8a9dacdb74ad5899b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f5fa79f0260e461c952301f904101f79', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 883.969719] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 883.969719] env[62914]: value = "task-1352648" [ 883.969719] env[62914]: _type = "Task" [ 883.969719] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.983043] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352648, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.069455] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e586d02-6a36-46e7-b852-f1bb236dbfa6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.076293] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2300d2a4-9786-456b-9f5e-9662a12ec857 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.109980] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b087ce54-9abf-436d-8a12-94bdb931aabd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.117212] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c37891-0216-403e-8033-68c74565f9d5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.131497] env[62914]: DEBUG nova.compute.provider_tree [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.158957] env[62914]: INFO nova.compute.manager [-] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Took 1.30 seconds to deallocate network for instance. [ 884.199229] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352646, 'name': Rename_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.251884] env[62914]: DEBUG nova.network.neutron [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Successfully created port: 9926578f-eff9-4aa1-b329-afaf16f5b7c0 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 884.275685] env[62914]: DEBUG oslo_vmware.api [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352632, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.303018] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352642, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.394174] env[62914]: DEBUG nova.compute.manager [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 884.480811] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352648, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066575} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.481099] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 884.481894] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b56e51-3b36-4d82-b947-c7b9f629c9bb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.506424] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] c9b94435-6546-464a-b3ba-c9e685157059/c9b94435-6546-464a-b3ba-c9e685157059.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.507099] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6f6dc11-0f9d-41ca-bcb3-f96501f6941a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.530018] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 884.530018] env[62914]: value = "task-1352649" [ 884.530018] env[62914]: _type = "Task" [ 884.530018] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.538437] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352649, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.638512] env[62914]: DEBUG nova.scheduler.client.report [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.665442] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.698181] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352646, 'name': Rename_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.750664] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Acquiring lock "93d420a1-6d8f-4919-a42f-55aebab853ae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.750945] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lock "93d420a1-6d8f-4919-a42f-55aebab853ae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.751194] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Acquiring lock "93d420a1-6d8f-4919-a42f-55aebab853ae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.751384] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lock "93d420a1-6d8f-4919-a42f-55aebab853ae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.752032] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lock "93d420a1-6d8f-4919-a42f-55aebab853ae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.753695] env[62914]: INFO nova.compute.manager [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Terminating instance [ 884.755630] env[62914]: DEBUG nova.compute.manager [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 884.755830] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 884.756661] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a5c12e-31ad-413a-b498-83d5d193252a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.764306] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 884.764568] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82cbc7ef-2a43-4a0f-a9f7-0d49c78166c3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.771162] env[62914]: DEBUG oslo_vmware.api [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 884.771162] env[62914]: value = "task-1352650" [ 884.771162] env[62914]: _type = "Task" [ 884.771162] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.774631] env[62914]: DEBUG oslo_vmware.api [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352632, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.782160] env[62914]: DEBUG oslo_vmware.api [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352650, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.800292] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352642, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.039190] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352649, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.143954] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.760s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.146468] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.481s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.146775] env[62914]: DEBUG nova.objects.instance [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Lazy-loading 'resources' on Instance uuid 29680cfb-8df3-4ff7-b6ee-0982577cc9cf {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.186248] env[62914]: INFO nova.scheduler.client.report [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleted allocations for instance 43ee8948-8805-4d48-bd45-e93d2e2eb05d [ 885.199654] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352646, 'name': Rename_Task, 'duration_secs': 1.992196} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.200262] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 885.200262] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eb79ad77-34e2-49c4-a911-78370ea017cb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.208183] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Waiting for the task: (returnval){ [ 885.208183] env[62914]: value = "task-1352651" [ 885.208183] env[62914]: _type = "Task" [ 885.208183] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.217783] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352651, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.285972] env[62914]: DEBUG oslo_vmware.api [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352632, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.292552] env[62914]: DEBUG oslo_vmware.api [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352650, 'name': PowerOffVM_Task, 'duration_secs': 0.194712} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.292913] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 885.293112] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 885.296487] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fe192e3d-fa41-48a2-b509-395f03aaf912 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.303351] env[62914]: DEBUG oslo_vmware.api [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352642, 'name': PowerOnVM_Task, 'duration_secs': 2.388084} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.303610] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.303795] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12bb6507-66a9-4663-a52e-c47f5ef9eb40 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating instance '4bf76ce0-b5d2-4184-888c-d0ef39878356' progress to 100 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 885.374147] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 885.374772] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 885.374772] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Deleting the datastore file [datastore2] 93d420a1-6d8f-4919-a42f-55aebab853ae {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 885.375144] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-613d8623-6bd1-4a29-adb9-ddd0bad90e7f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.381966] env[62914]: DEBUG oslo_vmware.api [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for the task: (returnval){ [ 885.381966] env[62914]: value = "task-1352653" [ 885.381966] env[62914]: _type = "Task" [ 885.381966] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.392837] env[62914]: DEBUG oslo_vmware.api [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352653, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.405021] env[62914]: DEBUG nova.compute.manager [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 885.435446] env[62914]: DEBUG nova.virt.hardware [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 885.435751] env[62914]: DEBUG nova.virt.hardware [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 885.435917] env[62914]: DEBUG nova.virt.hardware [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.436134] env[62914]: DEBUG nova.virt.hardware [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 885.436304] env[62914]: DEBUG nova.virt.hardware [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.436461] env[62914]: DEBUG nova.virt.hardware [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 885.436698] env[62914]: DEBUG nova.virt.hardware [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 885.436875] env[62914]: DEBUG nova.virt.hardware [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 885.437074] env[62914]: DEBUG nova.virt.hardware [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 885.437252] env[62914]: DEBUG nova.virt.hardware [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 885.437431] env[62914]: DEBUG nova.virt.hardware [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 885.438600] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-187cd055-f23f-4f22-a508-39bfcbf6b889 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.447164] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f97f6ca4-2f25-4f1e-8790-5b5856481013 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.539811] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352649, 'name': ReconfigVM_Task, 'duration_secs': 0.907541} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.540255] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Reconfigured VM instance instance-0000004f to attach disk [datastore2] c9b94435-6546-464a-b3ba-c9e685157059/c9b94435-6546-464a-b3ba-c9e685157059.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.540956] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a56c1a5e-bd5b-4797-9fd5-68ed54c017b7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.547871] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 885.547871] env[62914]: value = "task-1352654" [ 885.547871] env[62914]: _type = "Task" [ 885.547871] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.557270] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352654, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.697784] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5669a4a6-a671-4aa1-8143-0a429300dbca tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "43ee8948-8805-4d48-bd45-e93d2e2eb05d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.904s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.719465] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352651, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.778940] env[62914]: DEBUG oslo_vmware.api [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352632, 'name': ReconfigVM_Task, 'duration_secs': 5.842529} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.782120] env[62914]: DEBUG oslo_concurrency.lockutils [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.782354] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Reconfigured VM to detach interface {{(pid=62914) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 885.857806] env[62914]: DEBUG nova.compute.manager [req-83735418-abc7-4c09-ae90-53b721f75870 req-8c98259a-d51e-4476-aa45-f2b842382585 service nova] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Received event network-vif-plugged-9926578f-eff9-4aa1-b329-afaf16f5b7c0 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.857806] env[62914]: DEBUG oslo_concurrency.lockutils [req-83735418-abc7-4c09-ae90-53b721f75870 req-8c98259a-d51e-4476-aa45-f2b842382585 service nova] Acquiring lock "aec39e18-7796-4be3-af74-478df3a78f8f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.858060] env[62914]: DEBUG oslo_concurrency.lockutils [req-83735418-abc7-4c09-ae90-53b721f75870 req-8c98259a-d51e-4476-aa45-f2b842382585 service nova] Lock "aec39e18-7796-4be3-af74-478df3a78f8f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.858345] env[62914]: DEBUG oslo_concurrency.lockutils [req-83735418-abc7-4c09-ae90-53b721f75870 req-8c98259a-d51e-4476-aa45-f2b842382585 service nova] Lock "aec39e18-7796-4be3-af74-478df3a78f8f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.858554] env[62914]: DEBUG nova.compute.manager [req-83735418-abc7-4c09-ae90-53b721f75870 req-8c98259a-d51e-4476-aa45-f2b842382585 service nova] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] No waiting events found dispatching network-vif-plugged-9926578f-eff9-4aa1-b329-afaf16f5b7c0 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 885.858716] env[62914]: WARNING nova.compute.manager [req-83735418-abc7-4c09-ae90-53b721f75870 req-8c98259a-d51e-4476-aa45-f2b842382585 service nova] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Received unexpected event network-vif-plugged-9926578f-eff9-4aa1-b329-afaf16f5b7c0 for instance with vm_state building and task_state spawning. [ 885.862570] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a41c4dd-a4a5-45e0-b94c-d19f6beaa24a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.872775] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89495e90-3c67-45c9-a010-5d08fbd1d35b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.917651] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab7c97d-a9e9-409d-a3db-039fb543004d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.924918] env[62914]: DEBUG oslo_vmware.api [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Task: {'id': task-1352653, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176533} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.925732] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 885.926083] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 885.926363] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 885.926609] env[62914]: INFO nova.compute.manager [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Took 1.17 seconds to destroy the instance on the hypervisor. [ 885.926990] env[62914]: DEBUG oslo.service.loopingcall [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.930319] env[62914]: DEBUG nova.compute.manager [-] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 885.930498] env[62914]: DEBUG nova.network.neutron [-] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 885.933771] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c68c10-1e5d-4e9a-bf47-36609201a3b3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.950268] env[62914]: DEBUG nova.compute.provider_tree [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.058539] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352654, 'name': Rename_Task, 'duration_secs': 0.154613} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.058860] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.059157] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b40131f-1876-44d5-a0cc-638dda563074 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.067031] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 886.067031] env[62914]: value = "task-1352655" [ 886.067031] env[62914]: _type = "Task" [ 886.067031] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.084360] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352655, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.089111] env[62914]: DEBUG nova.network.neutron [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Successfully updated port: 9926578f-eff9-4aa1-b329-afaf16f5b7c0 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.212155] env[62914]: DEBUG nova.compute.manager [req-604af889-144f-434d-8139-90e9821b0e06 req-31bd61cc-1ce8-4823-bcf1-cbc5456933b2 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Received event network-vif-deleted-69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.212419] env[62914]: INFO nova.compute.manager [req-604af889-144f-434d-8139-90e9821b0e06 req-31bd61cc-1ce8-4823-bcf1-cbc5456933b2 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Neutron deleted interface 69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6; detaching it from the instance and deleting it from the info cache [ 886.212499] env[62914]: DEBUG nova.network.neutron [req-604af889-144f-434d-8139-90e9821b0e06 req-31bd61cc-1ce8-4823-bcf1-cbc5456933b2 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Updating instance_info_cache with network_info: [{"id": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "address": "fa:16:3e:7a:f0:fa", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb12a55b6-c3", "ovs_interfaceid": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.223643] env[62914]: DEBUG oslo_vmware.api [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352651, 'name': PowerOnVM_Task, 'duration_secs': 0.545748} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.224252] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 886.224252] env[62914]: INFO nova.compute.manager [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Took 9.33 seconds to spawn the instance on the hypervisor. [ 886.224377] env[62914]: DEBUG nova.compute.manager [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 886.225061] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03e67ef-546f-4924-99aa-ec5a11f2ceb7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.456078] env[62914]: DEBUG nova.scheduler.client.report [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.580124] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352655, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.592074] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "refresh_cache-aec39e18-7796-4be3-af74-478df3a78f8f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.593058] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "refresh_cache-aec39e18-7796-4be3-af74-478df3a78f8f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.593058] env[62914]: DEBUG nova.network.neutron [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.721017] env[62914]: DEBUG oslo_concurrency.lockutils [req-604af889-144f-434d-8139-90e9821b0e06 req-31bd61cc-1ce8-4823-bcf1-cbc5456933b2 service nova] Acquiring lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.721017] env[62914]: DEBUG oslo_concurrency.lockutils [req-604af889-144f-434d-8139-90e9821b0e06 req-31bd61cc-1ce8-4823-bcf1-cbc5456933b2 service nova] Acquired lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.721017] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7678d814-de6e-47ae-892f-311f02869447 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.748939] env[62914]: DEBUG oslo_concurrency.lockutils [req-604af889-144f-434d-8139-90e9821b0e06 req-31bd61cc-1ce8-4823-bcf1-cbc5456933b2 service nova] Releasing lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.751495] env[62914]: WARNING nova.compute.manager [req-604af889-144f-434d-8139-90e9821b0e06 req-31bd61cc-1ce8-4823-bcf1-cbc5456933b2 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Detach interface failed, port_id=69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6, reason: No device with interface-id 69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6 exists on VM: nova.exception.NotFound: No device with interface-id 69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6 exists on VM [ 886.753199] env[62914]: INFO nova.compute.manager [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Took 20.45 seconds to build instance. [ 886.774470] env[62914]: DEBUG nova.network.neutron [-] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.964889] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.817s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.991761] env[62914]: INFO nova.scheduler.client.report [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Deleted allocations for instance 29680cfb-8df3-4ff7-b6ee-0982577cc9cf [ 887.078763] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352655, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.118519] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Acquiring lock "47e6f142-bdb2-4397-9b94-d962e68ee6bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.142032] env[62914]: DEBUG nova.network.neutron [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.184873] env[62914]: DEBUG oslo_concurrency.lockutils [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.185162] env[62914]: DEBUG oslo_concurrency.lockutils [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.185355] env[62914]: DEBUG nova.network.neutron [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 887.256819] env[62914]: DEBUG oslo_concurrency.lockutils [None req-94d0dfa4-74f6-40d1-912c-91fd19aaa176 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Lock "47e6f142-bdb2-4397-9b94-d962e68ee6bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.959s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.257211] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Lock "47e6f142-bdb2-4397-9b94-d962e68ee6bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.139s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.260763] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Acquiring lock "47e6f142-bdb2-4397-9b94-d962e68ee6bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.260763] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Lock "47e6f142-bdb2-4397-9b94-d962e68ee6bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.260763] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Lock "47e6f142-bdb2-4397-9b94-d962e68ee6bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.262623] env[62914]: INFO nova.compute.manager [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Terminating instance [ 887.266118] env[62914]: DEBUG nova.compute.manager [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.266118] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 887.266118] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e6dd53-e61f-4d50-981b-44010fae094e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.274684] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.274875] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1dfa9183-a9ef-4927-90ec-a13a713a86b5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.276824] env[62914]: INFO nova.compute.manager [-] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Took 1.35 seconds to deallocate network for instance. [ 887.285171] env[62914]: DEBUG oslo_vmware.api [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Waiting for the task: (returnval){ [ 887.285171] env[62914]: value = "task-1352657" [ 887.285171] env[62914]: _type = "Task" [ 887.285171] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.297683] env[62914]: DEBUG oslo_vmware.api [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352657, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.403179] env[62914]: DEBUG nova.network.neutron [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Updating instance_info_cache with network_info: [{"id": "9926578f-eff9-4aa1-b329-afaf16f5b7c0", "address": "fa:16:3e:10:c9:71", "network": {"id": "c9c30d64-cafe-48a5-b868-b20d90d8920d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-458611021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5fa79f0260e461c952301f904101f79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9926578f-ef", "ovs_interfaceid": "9926578f-eff9-4aa1-b329-afaf16f5b7c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.501308] env[62914]: DEBUG oslo_concurrency.lockutils [None req-2a3b1c9f-e7c8-4469-bb75-e8927ddeab22 tempest-ServerRescueTestJSONUnderV235-2095194637 tempest-ServerRescueTestJSONUnderV235-2095194637-project-member] Lock "29680cfb-8df3-4ff7-b6ee-0982577cc9cf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.836s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.584467] env[62914]: DEBUG oslo_vmware.api [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352655, 'name': PowerOnVM_Task, 'duration_secs': 1.172488} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.584467] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 887.584467] env[62914]: INFO nova.compute.manager [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Took 8.39 seconds to spawn the instance on the hypervisor. [ 887.584467] env[62914]: DEBUG nova.compute.manager [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.584765] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0c0acd-b6cb-4a37-8443-ddcbc6b282be {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.788086] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.788427] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.788658] env[62914]: DEBUG nova.objects.instance [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lazy-loading 'resources' on Instance uuid 93d420a1-6d8f-4919-a42f-55aebab853ae {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.803600] env[62914]: DEBUG oslo_vmware.api [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352657, 'name': PowerOffVM_Task, 'duration_secs': 0.265351} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.803883] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 887.804069] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 887.804337] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-01ba0b71-066c-48e2-928f-22868c7b238e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.839333] env[62914]: DEBUG oslo_concurrency.lockutils [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "4bf76ce0-b5d2-4184-888c-d0ef39878356" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.839524] env[62914]: DEBUG oslo_concurrency.lockutils [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "4bf76ce0-b5d2-4184-888c-d0ef39878356" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.839718] env[62914]: DEBUG nova.compute.manager [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Going to confirm migration 1 {{(pid=62914) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 887.856396] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.856629] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.885889] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 887.886135] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 887.886326] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Deleting the datastore file [datastore2] 47e6f142-bdb2-4397-9b94-d962e68ee6bb {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.886583] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a7fd0aa2-abf1-4628-82dc-ac7f0b863e0d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.893587] env[62914]: DEBUG oslo_vmware.api [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Waiting for the task: (returnval){ [ 887.893587] env[62914]: value = "task-1352659" [ 887.893587] env[62914]: _type = "Task" [ 887.893587] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.904873] env[62914]: DEBUG oslo_vmware.api [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352659, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.904925] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.905153] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.905361] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.905550] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.905720] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.907426] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "refresh_cache-aec39e18-7796-4be3-af74-478df3a78f8f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.907698] env[62914]: DEBUG nova.compute.manager [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Instance network_info: |[{"id": "9926578f-eff9-4aa1-b329-afaf16f5b7c0", "address": "fa:16:3e:10:c9:71", "network": {"id": "c9c30d64-cafe-48a5-b868-b20d90d8920d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-458611021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5fa79f0260e461c952301f904101f79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9926578f-ef", "ovs_interfaceid": "9926578f-eff9-4aa1-b329-afaf16f5b7c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 887.908372] env[62914]: INFO nova.compute.manager [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Terminating instance [ 887.912509] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:c9:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9926578f-eff9-4aa1-b329-afaf16f5b7c0', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 887.918692] env[62914]: DEBUG oslo.service.loopingcall [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.921473] env[62914]: DEBUG nova.compute.manager [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.921575] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 887.921824] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 887.922676] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbbdec56-6906-4b1c-918f-c225c4bd5bea {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.925216] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-264e199c-3bd2-4842-940e-5f97b95b26e9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.945999] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.947319] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52eb2c76-7000-4e25-bd0b-e796fc6144dc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.948845] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.948845] env[62914]: value = "task-1352660" [ 887.948845] env[62914]: _type = "Task" [ 887.948845] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.953148] env[62914]: DEBUG oslo_vmware.api [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 887.953148] env[62914]: value = "task-1352661" [ 887.953148] env[62914]: _type = "Task" [ 887.953148] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.959751] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352660, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.964419] env[62914]: DEBUG oslo_vmware.api [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352661, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.104545] env[62914]: DEBUG nova.network.neutron [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Updating instance_info_cache with network_info: [{"id": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "address": "fa:16:3e:7a:f0:fa", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb12a55b6-c3", "ovs_interfaceid": "b12a55b6-c396-44ab-a9ef-b183e97189e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.111517] env[62914]: INFO nova.compute.manager [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Took 20.69 seconds to build instance. [ 888.153213] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Acquiring lock "ef27919a-f205-49d9-88d8-b350fcf5cfac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.153490] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Lock "ef27919a-f205-49d9-88d8-b350fcf5cfac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.241772] env[62914]: DEBUG nova.compute.manager [req-eb5acb4d-e1b2-4b93-9b3a-563f64bbafef req-fc21b9be-9360-4cdd-9c39-0c719dbabd44 service nova] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Received event network-changed-9926578f-eff9-4aa1-b329-afaf16f5b7c0 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.241977] env[62914]: DEBUG nova.compute.manager [req-eb5acb4d-e1b2-4b93-9b3a-563f64bbafef req-fc21b9be-9360-4cdd-9c39-0c719dbabd44 service nova] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Refreshing instance network info cache due to event network-changed-9926578f-eff9-4aa1-b329-afaf16f5b7c0. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 888.242208] env[62914]: DEBUG oslo_concurrency.lockutils [req-eb5acb4d-e1b2-4b93-9b3a-563f64bbafef req-fc21b9be-9360-4cdd-9c39-0c719dbabd44 service nova] Acquiring lock "refresh_cache-aec39e18-7796-4be3-af74-478df3a78f8f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.242356] env[62914]: DEBUG oslo_concurrency.lockutils [req-eb5acb4d-e1b2-4b93-9b3a-563f64bbafef req-fc21b9be-9360-4cdd-9c39-0c719dbabd44 service nova] Acquired lock "refresh_cache-aec39e18-7796-4be3-af74-478df3a78f8f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.242670] env[62914]: DEBUG nova.network.neutron [req-eb5acb4d-e1b2-4b93-9b3a-563f64bbafef req-fc21b9be-9360-4cdd-9c39-0c719dbabd44 service nova] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Refreshing network info cache for port 9926578f-eff9-4aa1-b329-afaf16f5b7c0 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 888.351344] env[62914]: DEBUG nova.compute.manager [req-1d624c71-e069-4108-9c0a-820cd8f17338 req-18ef1c97-7119-4e69-b6bb-c28a07d2717e service nova] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Received event network-vif-deleted-9fadd240-af76-45c7-9f2b-f48b581f1362 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.360437] env[62914]: DEBUG nova.compute.utils [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 888.407786] env[62914]: DEBUG oslo_vmware.api [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Task: {'id': task-1352659, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155964} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.407786] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.407786] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 888.407786] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 888.407786] env[62914]: INFO nova.compute.manager [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Took 1.14 seconds to destroy the instance on the hypervisor. [ 888.407786] env[62914]: DEBUG oslo.service.loopingcall [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.408857] env[62914]: DEBUG nova.compute.manager [-] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 888.409114] env[62914]: DEBUG nova.network.neutron [-] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 888.462832] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352660, 'name': CreateVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.468433] env[62914]: DEBUG oslo_vmware.api [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352661, 'name': PowerOffVM_Task, 'duration_secs': 0.211719} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.470851] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 888.470851] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 888.470851] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bfdb5c81-4737-44a1-b435-d50ff573a7eb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.477094] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a3caf3-f467-4b56-aa31-abe34b197765 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.481956] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec2f5114-fcaa-47b9-b822-56a3aaaf7e6c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.514477] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca26a15-4aa0-4abb-8e66-7f442ca7cee7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.524290] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c61229-c870-4937-82a3-a2b9819957d4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.530612] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 888.531092] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 888.531394] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Deleting the datastore file [datastore2] 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 888.532119] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a46d4fc-c5cf-4e82-945d-d411894ace03 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.541708] env[62914]: DEBUG nova.compute.provider_tree [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.546978] env[62914]: DEBUG oslo_vmware.api [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 888.546978] env[62914]: value = "task-1352663" [ 888.546978] env[62914]: _type = "Task" [ 888.546978] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.555481] env[62914]: DEBUG oslo_vmware.api [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352663, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.607084] env[62914]: DEBUG oslo_concurrency.lockutils [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "refresh_cache-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.617009] env[62914]: DEBUG oslo_concurrency.lockutils [None req-86f76c1c-633e-46fc-850e-1a975ece9a19 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "c9b94435-6546-464a-b3ba-c9e685157059" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.200s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.655699] env[62914]: DEBUG nova.compute.manager [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 888.756458] env[62914]: DEBUG oslo_concurrency.lockutils [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.756458] env[62914]: DEBUG oslo_concurrency.lockutils [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.756544] env[62914]: DEBUG nova.network.neutron [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.760115] env[62914]: DEBUG nova.objects.instance [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lazy-loading 'info_cache' on Instance uuid 4bf76ce0-b5d2-4184-888c-d0ef39878356 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.828412] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "c9b94435-6546-464a-b3ba-c9e685157059" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.828743] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "c9b94435-6546-464a-b3ba-c9e685157059" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.829089] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "c9b94435-6546-464a-b3ba-c9e685157059-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.829309] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "c9b94435-6546-464a-b3ba-c9e685157059-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.829497] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "c9b94435-6546-464a-b3ba-c9e685157059-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.832383] env[62914]: INFO nova.compute.manager [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Terminating instance [ 888.836179] env[62914]: DEBUG nova.compute.manager [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 888.836358] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 888.837500] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb6f9bc-8791-4169-bf71-93b25fe7d576 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.850020] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 888.850020] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e85cb297-b8c1-4df6-a55c-38c7717278ca {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.862023] env[62914]: DEBUG oslo_vmware.api [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 888.862023] env[62914]: value = "task-1352664" [ 888.862023] env[62914]: _type = "Task" [ 888.862023] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.863733] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.874097] env[62914]: DEBUG oslo_vmware.api [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352664, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.967837] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352660, 'name': CreateVM_Task, 'duration_secs': 0.651047} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.970716] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 888.974727] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.974897] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.975250] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 888.975956] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-465cfceb-f1f4-47ff-acb9-a3d7c8f174cd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.983334] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 888.983334] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5262996c-fbc8-f813-0b7e-c36036dd5561" [ 888.983334] env[62914]: _type = "Task" [ 888.983334] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.994511] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5262996c-fbc8-f813-0b7e-c36036dd5561, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.012362] env[62914]: DEBUG nova.network.neutron [req-eb5acb4d-e1b2-4b93-9b3a-563f64bbafef req-fc21b9be-9360-4cdd-9c39-0c719dbabd44 service nova] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Updated VIF entry in instance network info cache for port 9926578f-eff9-4aa1-b329-afaf16f5b7c0. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 889.012362] env[62914]: DEBUG nova.network.neutron [req-eb5acb4d-e1b2-4b93-9b3a-563f64bbafef req-fc21b9be-9360-4cdd-9c39-0c719dbabd44 service nova] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Updating instance_info_cache with network_info: [{"id": "9926578f-eff9-4aa1-b329-afaf16f5b7c0", "address": "fa:16:3e:10:c9:71", "network": {"id": "c9c30d64-cafe-48a5-b868-b20d90d8920d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-458611021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5fa79f0260e461c952301f904101f79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9926578f-ef", "ovs_interfaceid": "9926578f-eff9-4aa1-b329-afaf16f5b7c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.045714] env[62914]: DEBUG nova.scheduler.client.report [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 889.059666] env[62914]: DEBUG oslo_vmware.api [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352663, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14499} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.060350] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.061554] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 889.061797] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 889.062021] env[62914]: INFO nova.compute.manager [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Took 1.14 seconds to destroy the instance on the hypervisor. [ 889.062286] env[62914]: DEBUG oslo.service.loopingcall [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.062882] env[62914]: DEBUG nova.compute.manager [-] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 889.063342] env[62914]: DEBUG nova.network.neutron [-] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 889.111924] env[62914]: DEBUG oslo_concurrency.lockutils [None req-87cc73be-2979-4051-b2ea-9fa35874bc22 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07-69f9fa9f-a0ec-46dd-a9e3-0fc3cc0381c6" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.961s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.184207] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.336015] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Volume attach. Driver type: vmdk {{(pid=62914) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 889.336015] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288219', 'volume_id': '3b0294c0-2d37-450a-a517-8c71f44b449e', 'name': 'volume-3b0294c0-2d37-450a-a517-8c71f44b449e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '22de53a6-096b-4e8e-af91-f78588ea4c60', 'attached_at': '', 'detached_at': '', 'volume_id': '3b0294c0-2d37-450a-a517-8c71f44b449e', 'serial': '3b0294c0-2d37-450a-a517-8c71f44b449e'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 889.336015] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a87b40-5731-433e-a8c3-7483af6b76fb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.357889] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f59dd6-3480-47e6-90e6-64ea166eaa2e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.396738] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] volume-3b0294c0-2d37-450a-a517-8c71f44b449e/volume-3b0294c0-2d37-450a-a517-8c71f44b449e.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.401173] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84bed0d8-0f51-4673-886f-9b5c47725edc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.424423] env[62914]: DEBUG oslo_vmware.api [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352664, 'name': PowerOffVM_Task, 'duration_secs': 0.372658} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.424423] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 889.424502] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 889.424834] env[62914]: DEBUG oslo_vmware.api [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 889.424834] env[62914]: value = "task-1352665" [ 889.424834] env[62914]: _type = "Task" [ 889.424834] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.425015] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a436498-cf73-4619-95d2-c165e036f734 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.435403] env[62914]: DEBUG oslo_vmware.api [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352665, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.494523] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5262996c-fbc8-f813-0b7e-c36036dd5561, 'name': SearchDatastore_Task, 'duration_secs': 0.022477} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.495040] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.497013] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 889.497013] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.497013] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.497013] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 889.497013] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce9fdc94-47b2-47ef-b402-aaae43d14fe6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.504766] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 889.504766] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 889.504766] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleting the datastore file [datastore2] c9b94435-6546-464a-b3ba-c9e685157059 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.505054] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-121e0936-eb9c-4ca9-bf82-545ea445827c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.508749] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 889.509320] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 889.514479] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c59507b4-1fb3-476c-ae2d-63a15937ae7e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.519205] env[62914]: DEBUG oslo_concurrency.lockutils [req-eb5acb4d-e1b2-4b93-9b3a-563f64bbafef req-fc21b9be-9360-4cdd-9c39-0c719dbabd44 service nova] Releasing lock "refresh_cache-aec39e18-7796-4be3-af74-478df3a78f8f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.519801] env[62914]: DEBUG oslo_vmware.api [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 889.519801] env[62914]: value = "task-1352667" [ 889.519801] env[62914]: _type = "Task" [ 889.519801] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.526495] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 889.526495] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52eb3919-6396-80ca-5a08-bee4b792c6e6" [ 889.526495] env[62914]: _type = "Task" [ 889.526495] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.534047] env[62914]: DEBUG oslo_vmware.api [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.539453] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52eb3919-6396-80ca-5a08-bee4b792c6e6, 'name': SearchDatastore_Task, 'duration_secs': 0.009659} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.540295] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9f1192c-dd32-469e-8842-7a6444b68bb5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.545435] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 889.545435] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f4163d-adaa-6458-5bea-282dc9dd4a10" [ 889.545435] env[62914]: _type = "Task" [ 889.545435] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.556076] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.767s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.557449] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f4163d-adaa-6458-5bea-282dc9dd4a10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.557904] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.374s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.559397] env[62914]: INFO nova.compute.claims [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.572624] env[62914]: DEBUG nova.network.neutron [-] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.582320] env[62914]: INFO nova.scheduler.client.report [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Deleted allocations for instance 93d420a1-6d8f-4919-a42f-55aebab853ae [ 889.938131] env[62914]: DEBUG oslo_vmware.api [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352665, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.946999] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.947354] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.947563] env[62914]: INFO nova.compute.manager [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Attaching volume a3d96dcf-656d-4c56-a143-98acaae4b85f to /dev/sdb [ 890.000541] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628934b2-f30c-44c6-b2ed-0c451b63839d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.007696] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92398ad3-172d-42fb-bc78-3e46961677d4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.022103] env[62914]: DEBUG nova.network.neutron [-] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.023664] env[62914]: DEBUG nova.virt.block_device [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Updating existing volume attachment record: d194fadd-a77f-4561-a8a2-ce919bbff57f {{(pid=62914) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 890.036315] env[62914]: DEBUG oslo_vmware.api [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352667, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180663} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.036315] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.036315] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 890.036315] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 890.036315] env[62914]: INFO nova.compute.manager [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Took 1.20 seconds to destroy the instance on the hypervisor. [ 890.036578] env[62914]: DEBUG oslo.service.loopingcall [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.036765] env[62914]: DEBUG nova.compute.manager [-] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 890.036864] env[62914]: DEBUG nova.network.neutron [-] [instance: c9b94435-6546-464a-b3ba-c9e685157059] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 890.062773] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f4163d-adaa-6458-5bea-282dc9dd4a10, 'name': SearchDatastore_Task, 'duration_secs': 0.010174} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.063785] env[62914]: DEBUG nova.network.neutron [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating instance_info_cache with network_info: [{"id": "189c8622-547d-4da9-94da-564df3bc382d", "address": "fa:16:3e:b6:1a:f0", "network": {"id": "21b6ff73-feab-478c-848a-313ea0acbfc2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1273337660-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "619be7b3bf1445b68ba7adefc98f8782", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap189c8622-54", "ovs_interfaceid": "189c8622-547d-4da9-94da-564df3bc382d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.064937] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.065223] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] aec39e18-7796-4be3-af74-478df3a78f8f/aec39e18-7796-4be3-af74-478df3a78f8f.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 890.067957] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6802b402-2be7-45dd-881c-1add4da9bbfc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.076046] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 890.076046] env[62914]: value = "task-1352668" [ 890.076046] env[62914]: _type = "Task" [ 890.076046] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.079934] env[62914]: INFO nova.compute.manager [-] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Took 1.67 seconds to deallocate network for instance. [ 890.091899] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352668, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.091899] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7886d34-1dd8-4acf-a929-1199bc5c3fc2 tempest-ImagesOneServerTestJSON-946893260 tempest-ImagesOneServerTestJSON-946893260-project-member] Lock "93d420a1-6d8f-4919-a42f-55aebab853ae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.341s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.278184] env[62914]: DEBUG nova.compute.manager [req-a8d0c77f-2ebc-4f9d-a6c7-81ad7834df64 req-b335c63f-72ab-4541-92ce-48618dac0445 service nova] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Received event network-vif-deleted-b12a55b6-c396-44ab-a9ef-b183e97189e3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.388159] env[62914]: DEBUG nova.compute.manager [req-37d9d28c-6a8c-450b-aed9-3048d55ec62c req-0d7c75b2-c958-4aea-be25-fb123e6d8e9a service nova] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Received event network-vif-deleted-b705ac25-181c-430d-9661-805be91bb2a1 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.441292] env[62914]: DEBUG oslo_vmware.api [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352665, 'name': ReconfigVM_Task, 'duration_secs': 0.544319} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.441410] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Reconfigured VM instance instance-0000004d to attach disk [datastore1] volume-3b0294c0-2d37-450a-a517-8c71f44b449e/volume-3b0294c0-2d37-450a-a517-8c71f44b449e.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.447026] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8c70872-bc8d-4673-bd3a-baa0b61bf677 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.466759] env[62914]: DEBUG oslo_vmware.api [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 890.466759] env[62914]: value = "task-1352672" [ 890.466759] env[62914]: _type = "Task" [ 890.466759] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.478125] env[62914]: DEBUG oslo_vmware.api [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352672, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.528304] env[62914]: INFO nova.compute.manager [-] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Took 1.47 seconds to deallocate network for instance. [ 890.568898] env[62914]: DEBUG oslo_concurrency.lockutils [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "refresh_cache-4bf76ce0-b5d2-4184-888c-d0ef39878356" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.569249] env[62914]: DEBUG nova.objects.instance [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lazy-loading 'migration_context' on Instance uuid 4bf76ce0-b5d2-4184-888c-d0ef39878356 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 890.592920] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352668, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.593803] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.801365] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e780b65-372d-474b-827c-22c373a2a770 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.809693] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd19402-d463-4d8c-9162-aeef4878b83f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.841151] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b58dda-5e7b-43b5-9878-9e5c4a3f34d0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.854863] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a1a72cc-4873-4979-b4fe-ba9dac905906 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.870932] env[62914]: DEBUG nova.compute.provider_tree [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.914033] env[62914]: DEBUG nova.network.neutron [-] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.981192] env[62914]: DEBUG oslo_vmware.api [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352672, 'name': ReconfigVM_Task, 'duration_secs': 0.169658} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.981192] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288219', 'volume_id': '3b0294c0-2d37-450a-a517-8c71f44b449e', 'name': 'volume-3b0294c0-2d37-450a-a517-8c71f44b449e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '22de53a6-096b-4e8e-af91-f78588ea4c60', 'attached_at': '', 'detached_at': '', 'volume_id': '3b0294c0-2d37-450a-a517-8c71f44b449e', 'serial': '3b0294c0-2d37-450a-a517-8c71f44b449e'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 891.037704] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.072460] env[62914]: DEBUG nova.objects.base [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Object Instance<4bf76ce0-b5d2-4184-888c-d0ef39878356> lazy-loaded attributes: info_cache,migration_context {{(pid=62914) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 891.073459] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecefe50-874d-4a70-b559-63f4b204da56 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.086664] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352668, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681884} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.100172] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] aec39e18-7796-4be3-af74-478df3a78f8f/aec39e18-7796-4be3-af74-478df3a78f8f.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 891.100172] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 891.100782] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b9e73635-43c6-4098-9273-0cbfdaa0e62c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.102764] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2c59938-bf9b-4a59-9a4b-38ca6ca9ba15 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.107992] env[62914]: DEBUG oslo_vmware.api [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 891.107992] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ed0843-fecc-3f10-6469-44b243dac05b" [ 891.107992] env[62914]: _type = "Task" [ 891.107992] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.112109] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 891.112109] env[62914]: value = "task-1352673" [ 891.112109] env[62914]: _type = "Task" [ 891.112109] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.119471] env[62914]: DEBUG oslo_vmware.api [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ed0843-fecc-3f10-6469-44b243dac05b, 'name': SearchDatastore_Task, 'duration_secs': 0.007249} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.120103] env[62914]: DEBUG oslo_concurrency.lockutils [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.124150] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352673, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.377019] env[62914]: DEBUG nova.scheduler.client.report [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.417834] env[62914]: INFO nova.compute.manager [-] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Took 1.38 seconds to deallocate network for instance. [ 891.628969] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352673, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067641} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.628969] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 891.633173] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad1650b-a2d0-4c75-b0cc-16fc4826e225 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.655455] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] aec39e18-7796-4be3-af74-478df3a78f8f/aec39e18-7796-4be3-af74-478df3a78f8f.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.657067] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ed053c8-193f-47d4-bb26-fcd68a028912 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.677057] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 891.677057] env[62914]: value = "task-1352674" [ 891.677057] env[62914]: _type = "Task" [ 891.677057] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.685059] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352674, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.880650] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.881314] env[62914]: DEBUG nova.compute.manager [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 891.883734] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.291s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.883988] env[62914]: DEBUG nova.objects.instance [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Lazy-loading 'resources' on Instance uuid 47e6f142-bdb2-4397-9b94-d962e68ee6bb {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.927026] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.102023] env[62914]: DEBUG nova.objects.instance [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lazy-loading 'flavor' on Instance uuid 22de53a6-096b-4e8e-af91-f78588ea4c60 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.188422] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352674, 'name': ReconfigVM_Task, 'duration_secs': 0.259523} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.188775] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Reconfigured VM instance instance-00000050 to attach disk [datastore1] aec39e18-7796-4be3-af74-478df3a78f8f/aec39e18-7796-4be3-af74-478df3a78f8f.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.189568] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cb0124e3-dc65-4df6-88a8-4ed704c28b15 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.196339] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 892.196339] env[62914]: value = "task-1352675" [ 892.196339] env[62914]: _type = "Task" [ 892.196339] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.204700] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352675, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.362373] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "22de53a6-096b-4e8e-af91-f78588ea4c60" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.390161] env[62914]: DEBUG nova.compute.utils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.391730] env[62914]: DEBUG nova.compute.manager [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 892.391865] env[62914]: DEBUG nova.network.neutron [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 892.418080] env[62914]: DEBUG nova.compute.manager [req-65a2edf8-0bc0-4dc7-9501-8a1ff5960c75 req-45436456-edab-48f4-9c63-71ec1e22c136 service nova] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Received event network-vif-deleted-b5e84893-0b61-4ca8-b73b-1102b6477cf8 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.439971] env[62914]: DEBUG nova.policy [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3af1c41f79c04c59a66048ca614e6787', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0194cffdb3264565b84a6bdae38a1387', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 892.529219] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232b032a-fe37-487a-abde-50557db620d5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.537341] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f603d04e-6151-4908-902a-90eb207506bf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.569237] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f17349b-3708-4bb6-a44c-b5568c14a616 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.576762] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98577ab-3516-49ff-a483-6bded848fe7a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.591581] env[62914]: DEBUG nova.compute.provider_tree [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.606203] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f105cf64-80ff-4136-8af8-62d3a5f92149 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "22de53a6-096b-4e8e-af91-f78588ea4c60" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.917s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.607214] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "22de53a6-096b-4e8e-af91-f78588ea4c60" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.245s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.607358] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "22de53a6-096b-4e8e-af91-f78588ea4c60-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.607536] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "22de53a6-096b-4e8e-af91-f78588ea4c60-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.607709] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "22de53a6-096b-4e8e-af91-f78588ea4c60-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.610125] env[62914]: INFO nova.compute.manager [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Terminating instance [ 892.612596] env[62914]: DEBUG nova.compute.manager [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 892.612938] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 892.613015] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f5663e2-3beb-4440-a8d1-f23d07709939 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.619942] env[62914]: DEBUG oslo_vmware.api [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 892.619942] env[62914]: value = "task-1352677" [ 892.619942] env[62914]: _type = "Task" [ 892.619942] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.631561] env[62914]: DEBUG oslo_vmware.api [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352677, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.711160] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352675, 'name': Rename_Task, 'duration_secs': 0.171813} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.711352] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 892.711864] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-89edb5ec-52c8-4b4e-88a1-21eea7c32969 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.721643] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 892.721643] env[62914]: value = "task-1352678" [ 892.721643] env[62914]: _type = "Task" [ 892.721643] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.733605] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352678, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.788184] env[62914]: DEBUG nova.network.neutron [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Successfully created port: 52773cfb-b19a-47c3-aaae-7df97e237662 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.894998] env[62914]: DEBUG nova.compute.manager [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 893.094858] env[62914]: DEBUG nova.scheduler.client.report [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.130511] env[62914]: DEBUG oslo_vmware.api [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352677, 'name': PowerOffVM_Task, 'duration_secs': 0.156672} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.130850] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 893.131136] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Volume detach. Driver type: vmdk {{(pid=62914) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 893.131385] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288219', 'volume_id': '3b0294c0-2d37-450a-a517-8c71f44b449e', 'name': 'volume-3b0294c0-2d37-450a-a517-8c71f44b449e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '22de53a6-096b-4e8e-af91-f78588ea4c60', 'attached_at': '', 'detached_at': '', 'volume_id': '3b0294c0-2d37-450a-a517-8c71f44b449e', 'serial': '3b0294c0-2d37-450a-a517-8c71f44b449e'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 893.132227] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed3430c7-db6d-4c61-b464-cff72f2ad093 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.155298] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a736db3b-8160-4832-8851-19f446f04c98 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.165451] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a861b9-4436-4bff-a2cb-53ce1ec02878 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.190948] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2170e67-d25f-475d-a1af-be3ae1adf37e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.216117] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] The volume has not been displaced from its original location: [datastore1] volume-3b0294c0-2d37-450a-a517-8c71f44b449e/volume-3b0294c0-2d37-450a-a517-8c71f44b449e.vmdk. No consolidation needed. {{(pid=62914) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 893.222976] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 893.223096] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af4cf192-f49f-417a-9b7c-1ae85c7971e5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.246468] env[62914]: DEBUG oslo_vmware.api [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352678, 'name': PowerOnVM_Task, 'duration_secs': 0.478492} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.247719] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 893.247939] env[62914]: INFO nova.compute.manager [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Took 7.84 seconds to spawn the instance on the hypervisor. [ 893.248136] env[62914]: DEBUG nova.compute.manager [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 893.248504] env[62914]: DEBUG oslo_vmware.api [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 893.248504] env[62914]: value = "task-1352679" [ 893.248504] env[62914]: _type = "Task" [ 893.248504] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.249171] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a964a8-a6fd-437a-b70c-555401887c42 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.263209] env[62914]: DEBUG oslo_vmware.api [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352679, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.600919] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.717s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.603983] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.566s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.604284] env[62914]: DEBUG nova.objects.instance [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'resources' on Instance uuid 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.633751] env[62914]: INFO nova.scheduler.client.report [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Deleted allocations for instance 47e6f142-bdb2-4397-9b94-d962e68ee6bb [ 893.760848] env[62914]: DEBUG oslo_vmware.api [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352679, 'name': ReconfigVM_Task, 'duration_secs': 0.473722} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.761452] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 893.770447] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1dfb0b4b-ff38-407a-a2cf-b5dce30d7a2e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.782542] env[62914]: INFO nova.compute.manager [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Took 13.10 seconds to build instance. [ 893.791205] env[62914]: DEBUG oslo_vmware.api [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 893.791205] env[62914]: value = "task-1352680" [ 893.791205] env[62914]: _type = "Task" [ 893.791205] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.801619] env[62914]: DEBUG oslo_vmware.api [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352680, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.889017] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "c7cf71d7-b823-471f-949f-9ef2b539d972" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.889916] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.889916] env[62914]: INFO nova.compute.manager [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Shelving [ 893.903422] env[62914]: DEBUG nova.compute.manager [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 893.929469] env[62914]: DEBUG nova.virt.hardware [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.929705] env[62914]: DEBUG nova.virt.hardware [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.929862] env[62914]: DEBUG nova.virt.hardware [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.930056] env[62914]: DEBUG nova.virt.hardware [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.930202] env[62914]: DEBUG nova.virt.hardware [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.930383] env[62914]: DEBUG nova.virt.hardware [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.930627] env[62914]: DEBUG nova.virt.hardware [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.930789] env[62914]: DEBUG nova.virt.hardware [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.930953] env[62914]: DEBUG nova.virt.hardware [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.931130] env[62914]: DEBUG nova.virt.hardware [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.931305] env[62914]: DEBUG nova.virt.hardware [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.933023] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95fd109a-697b-4034-b964-2c58ef564485 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.940546] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349e52a3-ff04-41a7-a35a-3caa513a7524 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.142849] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ac8c8efc-18f3-4118-b502-bf5d03ef7fe1 tempest-InstanceActionsV221TestJSON-1452259966 tempest-InstanceActionsV221TestJSON-1452259966-project-member] Lock "47e6f142-bdb2-4397-9b94-d962e68ee6bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.886s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.876090] env[62914]: DEBUG nova.network.neutron [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Successfully updated port: 52773cfb-b19a-47c3-aaae-7df97e237662 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.880079] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Volume attach. Driver type: vmdk {{(pid=62914) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 894.880304] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288222', 'volume_id': 'a3d96dcf-656d-4c56-a143-98acaae4b85f', 'name': 'volume-a3d96dcf-656d-4c56-a143-98acaae4b85f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1fe15423-f8be-4763-b55e-2cbb383ff01d', 'attached_at': '', 'detached_at': '', 'volume_id': 'a3d96dcf-656d-4c56-a143-98acaae4b85f', 'serial': 'a3d96dcf-656d-4c56-a143-98acaae4b85f'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 894.881590] env[62914]: DEBUG oslo_concurrency.lockutils [None req-98672863-eee0-4423-99cb-3d098ebf7b1f tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "aec39e18-7796-4be3-af74-478df3a78f8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.212s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.885660] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff23058e-bd88-4b6e-a419-b4888d6f18fc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.888895] env[62914]: DEBUG nova.compute.manager [req-9eb6a014-241e-4f2e-b323-6ab0300f4cc4 req-7b253400-789f-4760-9077-56c8d6ddecd3 service nova] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Received event network-vif-plugged-52773cfb-b19a-47c3-aaae-7df97e237662 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.889106] env[62914]: DEBUG oslo_concurrency.lockutils [req-9eb6a014-241e-4f2e-b323-6ab0300f4cc4 req-7b253400-789f-4760-9077-56c8d6ddecd3 service nova] Acquiring lock "ef27919a-f205-49d9-88d8-b350fcf5cfac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.889314] env[62914]: DEBUG oslo_concurrency.lockutils [req-9eb6a014-241e-4f2e-b323-6ab0300f4cc4 req-7b253400-789f-4760-9077-56c8d6ddecd3 service nova] Lock "ef27919a-f205-49d9-88d8-b350fcf5cfac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.889499] env[62914]: DEBUG oslo_concurrency.lockutils [req-9eb6a014-241e-4f2e-b323-6ab0300f4cc4 req-7b253400-789f-4760-9077-56c8d6ddecd3 service nova] Lock "ef27919a-f205-49d9-88d8-b350fcf5cfac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.889682] env[62914]: DEBUG nova.compute.manager [req-9eb6a014-241e-4f2e-b323-6ab0300f4cc4 req-7b253400-789f-4760-9077-56c8d6ddecd3 service nova] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] No waiting events found dispatching network-vif-plugged-52773cfb-b19a-47c3-aaae-7df97e237662 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 894.889849] env[62914]: WARNING nova.compute.manager [req-9eb6a014-241e-4f2e-b323-6ab0300f4cc4 req-7b253400-789f-4760-9077-56c8d6ddecd3 service nova] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Received unexpected event network-vif-plugged-52773cfb-b19a-47c3-aaae-7df97e237662 for instance with vm_state building and task_state spawning. [ 894.895889] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 894.897008] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8cc0f671-5022-427e-b506-186a49bbb370 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.899856] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa766394-0416-4704-9e99-dba95d68968a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.916397] env[62914]: DEBUG oslo_vmware.api [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352680, 'name': ReconfigVM_Task, 'duration_secs': 0.297211} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.920117] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288219', 'volume_id': '3b0294c0-2d37-450a-a517-8c71f44b449e', 'name': 'volume-3b0294c0-2d37-450a-a517-8c71f44b449e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '22de53a6-096b-4e8e-af91-f78588ea4c60', 'attached_at': '', 'detached_at': '', 'volume_id': '3b0294c0-2d37-450a-a517-8c71f44b449e', 'serial': '3b0294c0-2d37-450a-a517-8c71f44b449e'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 894.920528] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 894.921287] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ebbe34-0c61-4d5e-b406-d07b164d3d0c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.927077] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-953284ca-a193-49bb-8bee-7aa6ee26f973 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.930979] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 894.930979] env[62914]: value = "task-1352681" [ 894.930979] env[62914]: _type = "Task" [ 894.930979] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.950686] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7a9585b7-3cda-4813-aca1-10d99930593b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Suspending the VM {{(pid=62914) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 894.951052] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 894.959893] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] volume-a3d96dcf-656d-4c56-a143-98acaae4b85f/volume-a3d96dcf-656d-4c56-a143-98acaae4b85f.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 894.962356] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-e97a1d85-1931-4926-a163-7f94de40cc28 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.963761] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-20eadec7-929f-4b70-ae5e-16dd9723915d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.965115] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0976e38-1701-4a0f-841f-bb1918ee2482 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.981301] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352681, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.988016] env[62914]: DEBUG oslo_vmware.api [None req-7a9585b7-3cda-4813-aca1-10d99930593b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 894.988016] env[62914]: value = "task-1352682" [ 894.988016] env[62914]: _type = "Task" [ 894.988016] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.988217] env[62914]: DEBUG oslo_vmware.api [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 894.988217] env[62914]: value = "task-1352684" [ 894.988217] env[62914]: _type = "Task" [ 894.988217] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.003567] env[62914]: DEBUG oslo_vmware.api [None req-7a9585b7-3cda-4813-aca1-10d99930593b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352682, 'name': SuspendVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.006886] env[62914]: DEBUG oslo_vmware.api [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352684, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.063310] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc92e95d-23b1-453e-8bff-3f7e28499dec {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.071606] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2acabb1-acc5-4934-933d-e7bdfdbee18e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.108848] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f87f66c2-46c1-4d1c-b1b7-b2db52e90caf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.112588] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.112822] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.113013] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleting the datastore file [datastore1] 22de53a6-096b-4e8e-af91-f78588ea4c60 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.113310] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1e0da7c8-7bd7-4431-a38e-8e6b2647e820 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.121457] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e49f50ff-df1d-4d19-a5ca-b9634582ab0f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.125430] env[62914]: DEBUG oslo_vmware.api [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 895.125430] env[62914]: value = "task-1352685" [ 895.125430] env[62914]: _type = "Task" [ 895.125430] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.140299] env[62914]: DEBUG nova.compute.provider_tree [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.143712] env[62914]: DEBUG oslo_vmware.api [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352685, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.391202] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Acquiring lock "refresh_cache-ef27919a-f205-49d9-88d8-b350fcf5cfac" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.391404] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Acquired lock "refresh_cache-ef27919a-f205-49d9-88d8-b350fcf5cfac" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.391592] env[62914]: DEBUG nova.network.neutron [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 895.445274] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352681, 'name': PowerOffVM_Task, 'duration_secs': 0.169821} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.445813] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.446692] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9791aed7-e18a-446c-b536-b8ef743a4c9e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.466325] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74bb7046-3588-47f3-861a-f1bb7ef20a4f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.505654] env[62914]: DEBUG oslo_vmware.api [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.513376] env[62914]: DEBUG oslo_vmware.api [None req-7a9585b7-3cda-4813-aca1-10d99930593b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352682, 'name': SuspendVM_Task} progress is 62%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.637440] env[62914]: DEBUG oslo_vmware.api [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352685, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.255877} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.637715] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.637909] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 895.638115] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 895.638316] env[62914]: INFO nova.compute.manager [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Took 3.03 seconds to destroy the instance on the hypervisor. [ 895.638603] env[62914]: DEBUG oslo.service.loopingcall [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.638831] env[62914]: DEBUG nova.compute.manager [-] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 895.638929] env[62914]: DEBUG nova.network.neutron [-] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 895.642347] env[62914]: DEBUG nova.scheduler.client.report [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.946456] env[62914]: DEBUG nova.network.neutron [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 895.980607] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Creating Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 895.981920] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b33a3952-9849-4fe7-b6fe-711dcd07adc7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.989556] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 895.989556] env[62914]: value = "task-1352686" [ 895.989556] env[62914]: _type = "Task" [ 895.989556] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.003683] env[62914]: DEBUG oslo_vmware.api [None req-7a9585b7-3cda-4813-aca1-10d99930593b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352682, 'name': SuspendVM_Task, 'duration_secs': 0.691516} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.007139] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7a9585b7-3cda-4813-aca1-10d99930593b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Suspended the VM {{(pid=62914) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 896.007370] env[62914]: DEBUG nova.compute.manager [None req-7a9585b7-3cda-4813-aca1-10d99930593b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.007659] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352686, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.010420] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf53938-5485-4883-bc61-5489a965ea8f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.015916] env[62914]: DEBUG oslo_vmware.api [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352684, 'name': ReconfigVM_Task, 'duration_secs': 0.661599} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.016479] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Reconfigured VM instance instance-00000047 to attach disk [datastore2] volume-a3d96dcf-656d-4c56-a143-98acaae4b85f/volume-a3d96dcf-656d-4c56-a143-98acaae4b85f.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 896.022288] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eaf5b1e7-fe40-4bb4-8ca3-34bd28267913 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.040022] env[62914]: DEBUG oslo_vmware.api [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 896.040022] env[62914]: value = "task-1352687" [ 896.040022] env[62914]: _type = "Task" [ 896.040022] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.051231] env[62914]: DEBUG oslo_vmware.api [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352687, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.118796] env[62914]: DEBUG nova.network.neutron [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Updating instance_info_cache with network_info: [{"id": "52773cfb-b19a-47c3-aaae-7df97e237662", "address": "fa:16:3e:17:ff:3e", "network": {"id": "0700bc9a-485a-4d13-97a8-2a9a4bf10420", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1862735526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0194cffdb3264565b84a6bdae38a1387", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52773cfb-b1", "ovs_interfaceid": "52773cfb-b19a-47c3-aaae-7df97e237662", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.147220] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.543s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.149890] env[62914]: DEBUG oslo_concurrency.lockutils [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 5.030s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.172641] env[62914]: INFO nova.scheduler.client.report [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Deleted allocations for instance 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07 [ 896.412383] env[62914]: DEBUG nova.network.neutron [-] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.442196] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Acquiring lock "f14f4ab4-5ef1-4f2c-ae97-ac924a446df0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.442421] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Lock "f14f4ab4-5ef1-4f2c-ae97-ac924a446df0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.485129] env[62914]: DEBUG nova.compute.manager [req-a28863f4-bd4a-40b5-8731-1d85ee7c5890 req-86c7ad5a-5c1b-480d-b554-4e5ea05f99cb service nova] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Received event network-changed-52773cfb-b19a-47c3-aaae-7df97e237662 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.485319] env[62914]: DEBUG nova.compute.manager [req-a28863f4-bd4a-40b5-8731-1d85ee7c5890 req-86c7ad5a-5c1b-480d-b554-4e5ea05f99cb service nova] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Refreshing instance network info cache due to event network-changed-52773cfb-b19a-47c3-aaae-7df97e237662. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.485513] env[62914]: DEBUG oslo_concurrency.lockutils [req-a28863f4-bd4a-40b5-8731-1d85ee7c5890 req-86c7ad5a-5c1b-480d-b554-4e5ea05f99cb service nova] Acquiring lock "refresh_cache-ef27919a-f205-49d9-88d8-b350fcf5cfac" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.503741] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352686, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.550601] env[62914]: DEBUG oslo_vmware.api [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352687, 'name': ReconfigVM_Task, 'duration_secs': 0.135684} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.550954] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288222', 'volume_id': 'a3d96dcf-656d-4c56-a143-98acaae4b85f', 'name': 'volume-a3d96dcf-656d-4c56-a143-98acaae4b85f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1fe15423-f8be-4763-b55e-2cbb383ff01d', 'attached_at': '', 'detached_at': '', 'volume_id': 'a3d96dcf-656d-4c56-a143-98acaae4b85f', 'serial': 'a3d96dcf-656d-4c56-a143-98acaae4b85f'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 896.624296] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Releasing lock "refresh_cache-ef27919a-f205-49d9-88d8-b350fcf5cfac" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.624760] env[62914]: DEBUG nova.compute.manager [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Instance network_info: |[{"id": "52773cfb-b19a-47c3-aaae-7df97e237662", "address": "fa:16:3e:17:ff:3e", "network": {"id": "0700bc9a-485a-4d13-97a8-2a9a4bf10420", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1862735526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0194cffdb3264565b84a6bdae38a1387", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52773cfb-b1", "ovs_interfaceid": "52773cfb-b19a-47c3-aaae-7df97e237662", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 896.625506] env[62914]: DEBUG oslo_concurrency.lockutils [req-a28863f4-bd4a-40b5-8731-1d85ee7c5890 req-86c7ad5a-5c1b-480d-b554-4e5ea05f99cb service nova] Acquired lock "refresh_cache-ef27919a-f205-49d9-88d8-b350fcf5cfac" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.625569] env[62914]: DEBUG nova.network.neutron [req-a28863f4-bd4a-40b5-8731-1d85ee7c5890 req-86c7ad5a-5c1b-480d-b554-4e5ea05f99cb service nova] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Refreshing network info cache for port 52773cfb-b19a-47c3-aaae-7df97e237662 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 896.626901] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:ff:3e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9b7e9e55-3210-4fae-9648-d87e76c3d931', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '52773cfb-b19a-47c3-aaae-7df97e237662', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 896.637381] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Creating folder: Project (0194cffdb3264565b84a6bdae38a1387). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 896.638432] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b6cbd6b6-5013-4ed8-810b-3b93ef52fa17 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.651080] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Created folder: Project (0194cffdb3264565b84a6bdae38a1387) in parent group-v288131. [ 896.651167] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Creating folder: Instances. Parent ref: group-v288223. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 896.651345] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1acd8fa3-928c-4ac7-b33a-d1e7ee8cec9d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.669191] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Created folder: Instances in parent group-v288223. [ 896.669463] env[62914]: DEBUG oslo.service.loopingcall [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.669693] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 896.669944] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5d6bc62-61cd-42b5-8084-4da6e78819b0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.692108] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f36b059-fb47-4b7b-876d-7b8f14786e8c tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.786s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.696145] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 896.696145] env[62914]: value = "task-1352690" [ 896.696145] env[62914]: _type = "Task" [ 896.696145] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.704583] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352690, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.850045] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d21e41b0-373d-4a31-9ad3-38da0fe07a9a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.857861] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3447f62-98bb-4ab7-b4f9-0a038e389d49 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.898392] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2092aa34-c515-41f2-adda-dc6254898996 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.908103] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19abe744-5a32-4a4e-9085-85362de4f195 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.923506] env[62914]: INFO nova.compute.manager [-] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Took 1.28 seconds to deallocate network for instance. [ 896.924031] env[62914]: DEBUG nova.compute.provider_tree [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.946018] env[62914]: DEBUG nova.compute.manager [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 897.005261] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352686, 'name': CreateSnapshot_Task, 'duration_secs': 0.863566} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.005549] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Created Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 897.006320] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e921032-f065-43a8-8e16-8345067b1219 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.206920] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352690, 'name': CreateVM_Task, 'duration_secs': 0.367209} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.207182] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 897.207925] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.208181] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.208580] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 897.208895] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30386a84-8aea-4b4c-acd8-5c08b694f764 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.214277] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Waiting for the task: (returnval){ [ 897.214277] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]520470fd-7bea-7600-2926-b9957cdd0f29" [ 897.214277] env[62914]: _type = "Task" [ 897.214277] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.222253] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]520470fd-7bea-7600-2926-b9957cdd0f29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.412099] env[62914]: DEBUG nova.network.neutron [req-a28863f4-bd4a-40b5-8731-1d85ee7c5890 req-86c7ad5a-5c1b-480d-b554-4e5ea05f99cb service nova] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Updated VIF entry in instance network info cache for port 52773cfb-b19a-47c3-aaae-7df97e237662. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 897.412486] env[62914]: DEBUG nova.network.neutron [req-a28863f4-bd4a-40b5-8731-1d85ee7c5890 req-86c7ad5a-5c1b-480d-b554-4e5ea05f99cb service nova] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Updating instance_info_cache with network_info: [{"id": "52773cfb-b19a-47c3-aaae-7df97e237662", "address": "fa:16:3e:17:ff:3e", "network": {"id": "0700bc9a-485a-4d13-97a8-2a9a4bf10420", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1862735526-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0194cffdb3264565b84a6bdae38a1387", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52773cfb-b1", "ovs_interfaceid": "52773cfb-b19a-47c3-aaae-7df97e237662", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.428548] env[62914]: DEBUG nova.scheduler.client.report [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.472683] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.479803] env[62914]: INFO nova.compute.manager [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Took 0.55 seconds to detach 1 volumes for instance. [ 897.531914] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Creating linked-clone VM from snapshot {{(pid=62914) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 897.534175] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0cba2cc9-aa29-4f5d-90d5-4dc561499a23 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.542377] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 897.542377] env[62914]: value = "task-1352691" [ 897.542377] env[62914]: _type = "Task" [ 897.542377] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.551174] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352691, 'name': CloneVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.598501] env[62914]: DEBUG nova.objects.instance [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lazy-loading 'flavor' on Instance uuid 1fe15423-f8be-4763-b55e-2cbb383ff01d {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.729878] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]520470fd-7bea-7600-2926-b9957cdd0f29, 'name': SearchDatastore_Task, 'duration_secs': 0.022359} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.729878] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.729878] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 897.729878] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.729878] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.729878] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 897.729878] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-633516cb-75fd-4548-9fe2-e7a57af0e06b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.738884] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 897.738884] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 897.739606] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c35924cd-4ec8-4d57-8b41-47dc7bff42e1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.745024] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Waiting for the task: (returnval){ [ 897.745024] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52600254-27f7-2068-4dd8-318876a8659f" [ 897.745024] env[62914]: _type = "Task" [ 897.745024] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.756125] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52600254-27f7-2068-4dd8-318876a8659f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.914871] env[62914]: DEBUG oslo_concurrency.lockutils [req-a28863f4-bd4a-40b5-8731-1d85ee7c5890 req-86c7ad5a-5c1b-480d-b554-4e5ea05f99cb service nova] Releasing lock "refresh_cache-ef27919a-f205-49d9-88d8-b350fcf5cfac" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.915186] env[62914]: DEBUG nova.compute.manager [req-a28863f4-bd4a-40b5-8731-1d85ee7c5890 req-86c7ad5a-5c1b-480d-b554-4e5ea05f99cb service nova] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Received event network-vif-deleted-02c6f8b8-aaaa-4b7e-979c-43ad85975569 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.986627] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.053029] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352691, 'name': CloneVM_Task} progress is 94%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.106045] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dc37e598-23f8-434c-b9b2-ab047eae2101 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.156s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.138979] env[62914]: DEBUG nova.compute.manager [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 898.139954] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba87696b-8bc5-45d3-9552-a27830787711 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.158403] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "ab036e1f-a62a-490a-818c-2e1256c781f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.158552] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "ab036e1f-a62a-490a-818c-2e1256c781f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.255357] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52600254-27f7-2068-4dd8-318876a8659f, 'name': SearchDatastore_Task, 'duration_secs': 0.0308} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.256404] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a6cda78-56cd-4c28-b7b7-83935a60ec7e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.262703] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Waiting for the task: (returnval){ [ 898.262703] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5202dae8-2a24-7dc7-4804-09b30cb9d375" [ 898.262703] env[62914]: _type = "Task" [ 898.262703] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.272330] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5202dae8-2a24-7dc7-4804-09b30cb9d375, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.442633] env[62914]: DEBUG oslo_concurrency.lockutils [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.293s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.445799] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.520s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.445799] env[62914]: DEBUG nova.objects.instance [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lazy-loading 'resources' on Instance uuid c9b94435-6546-464a-b3ba-c9e685157059 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.555623] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "b8aebca4-2724-419e-aaf7-d180e299a459" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.555931] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "b8aebca4-2724-419e-aaf7-d180e299a459" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.561020] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352691, 'name': CloneVM_Task} progress is 95%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.656347] env[62914]: INFO nova.compute.manager [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] instance snapshotting [ 898.656606] env[62914]: WARNING nova.compute.manager [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 898.660010] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b76425-38ae-4247-9d3b-8cd8886f5ade {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.663951] env[62914]: DEBUG nova.compute.manager [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 898.688345] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929204c4-b53d-4b52-95e6-577a490462bf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.772483] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5202dae8-2a24-7dc7-4804-09b30cb9d375, 'name': SearchDatastore_Task, 'duration_secs': 0.009883} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.772828] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.773105] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] ef27919a-f205-49d9-88d8-b350fcf5cfac/ef27919a-f205-49d9-88d8-b350fcf5cfac.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 898.773370] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-10106448-97d4-41f0-b7da-8054ea91960f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.780993] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Waiting for the task: (returnval){ [ 898.780993] env[62914]: value = "task-1352692" [ 898.780993] env[62914]: _type = "Task" [ 898.780993] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.793177] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352692, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.915593] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.915901] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.943780] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "1f7b1609-08e0-4439-aea0-b4a2ec808780" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.944065] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "1f7b1609-08e0-4439-aea0-b4a2ec808780" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.999553] env[62914]: INFO nova.scheduler.client.report [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Deleted allocation for migration 19c9417f-8ea9-4e58-b81e-a66228590190 [ 899.054641] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352691, 'name': CloneVM_Task, 'duration_secs': 1.321922} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.057457] env[62914]: INFO nova.virt.vmwareapi.vmops [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Created linked-clone VM from snapshot [ 899.058535] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a121d9-d38a-4572-a665-c7614680e096 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.061797] env[62914]: DEBUG nova.compute.manager [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 899.069382] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Uploading image 662656a7-d46d-4633-b79c-60fc2ed8c7dc {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 899.095382] env[62914]: DEBUG oslo_vmware.rw_handles [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 899.095382] env[62914]: value = "vm-288227" [ 899.095382] env[62914]: _type = "VirtualMachine" [ 899.095382] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 899.095864] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0547ca91-5c28-430a-8755-04a4238ea8ee {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.108187] env[62914]: DEBUG oslo_vmware.rw_handles [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lease: (returnval){ [ 899.108187] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]529898c3-2e81-e2cf-90b1-c12cbb78fe6e" [ 899.108187] env[62914]: _type = "HttpNfcLease" [ 899.108187] env[62914]: } obtained for exporting VM: (result){ [ 899.108187] env[62914]: value = "vm-288227" [ 899.108187] env[62914]: _type = "VirtualMachine" [ 899.108187] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 899.108620] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the lease: (returnval){ [ 899.108620] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]529898c3-2e81-e2cf-90b1-c12cbb78fe6e" [ 899.108620] env[62914]: _type = "HttpNfcLease" [ 899.108620] env[62914]: } to be ready. {{(pid=62914) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 899.118433] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 899.118433] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]529898c3-2e81-e2cf-90b1-c12cbb78fe6e" [ 899.118433] env[62914]: _type = "HttpNfcLease" [ 899.118433] env[62914]: } is initializing. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 899.138265] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f010704-c667-4a8f-820e-d34eed398fad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.148964] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3be28f7-6f95-47f6-aabf-9deade0ac7fb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.184486] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ec3f8a-a999-4055-b013-260fe4831735 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.193642] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd0877d-cf20-4be4-8b0f-03689fbec62f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.198817] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.201025] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Creating Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 899.209622] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d227adda-f39a-4495-a976-cce75247ca1a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.212339] env[62914]: DEBUG nova.compute.provider_tree [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.221740] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 899.221740] env[62914]: value = "task-1352694" [ 899.221740] env[62914]: _type = "Task" [ 899.221740] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.232316] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352694, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.291509] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352692, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507939} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.291789] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] ef27919a-f205-49d9-88d8-b350fcf5cfac/ef27919a-f205-49d9-88d8-b350fcf5cfac.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 899.292042] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 899.292283] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19c0f715-5299-43cf-bf99-7e4dcf7c7574 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.298403] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Waiting for the task: (returnval){ [ 899.298403] env[62914]: value = "task-1352695" [ 899.298403] env[62914]: _type = "Task" [ 899.298403] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.306662] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "1e527707-758c-4006-af83-9c739b9645ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.306892] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "1e527707-758c-4006-af83-9c739b9645ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.311702] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352695, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.419963] env[62914]: INFO nova.compute.manager [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Detaching volume a3d96dcf-656d-4c56-a143-98acaae4b85f [ 899.447619] env[62914]: DEBUG nova.compute.manager [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 899.461108] env[62914]: INFO nova.virt.block_device [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Attempting to driver detach volume a3d96dcf-656d-4c56-a143-98acaae4b85f from mountpoint /dev/sdb [ 899.461515] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Volume detach. Driver type: vmdk {{(pid=62914) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 899.462752] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288222', 'volume_id': 'a3d96dcf-656d-4c56-a143-98acaae4b85f', 'name': 'volume-a3d96dcf-656d-4c56-a143-98acaae4b85f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1fe15423-f8be-4763-b55e-2cbb383ff01d', 'attached_at': '', 'detached_at': '', 'volume_id': 'a3d96dcf-656d-4c56-a143-98acaae4b85f', 'serial': 'a3d96dcf-656d-4c56-a143-98acaae4b85f'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 899.463012] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-272e09c8-04ef-4ff1-8b29-a5eb33375d2f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.487670] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8015e6-09c5-48c7-a21a-9f9361d86b51 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.496815] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a32247fd-f157-4ba5-812e-8f85782bb6ac {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.519628] env[62914]: DEBUG oslo_concurrency.lockutils [None req-62965727-9ac0-4934-907d-997c9a0c55a5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "4bf76ce0-b5d2-4184-888c-d0ef39878356" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 11.680s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.521668] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef767139-f5a8-4d3c-b137-5ceaa61ce21d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.539328] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] The volume has not been displaced from its original location: [datastore2] volume-a3d96dcf-656d-4c56-a143-98acaae4b85f/volume-a3d96dcf-656d-4c56-a143-98acaae4b85f.vmdk. No consolidation needed. {{(pid=62914) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 899.544703] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Reconfiguring VM instance instance-00000047 to detach disk 2001 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 899.545403] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a970aac9-6d4c-4d77-947b-75354589851f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.566718] env[62914]: DEBUG oslo_vmware.api [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 899.566718] env[62914]: value = "task-1352696" [ 899.566718] env[62914]: _type = "Task" [ 899.566718] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.580012] env[62914]: DEBUG oslo_vmware.api [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352696, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.584724] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.621079] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 899.621079] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]529898c3-2e81-e2cf-90b1-c12cbb78fe6e" [ 899.621079] env[62914]: _type = "HttpNfcLease" [ 899.621079] env[62914]: } is ready. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 899.621079] env[62914]: DEBUG oslo_vmware.rw_handles [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 899.621079] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]529898c3-2e81-e2cf-90b1-c12cbb78fe6e" [ 899.621079] env[62914]: _type = "HttpNfcLease" [ 899.621079] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 899.621790] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6de3bf-d9c0-4bd0-8d14-0cb355e8c67f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.629693] env[62914]: DEBUG oslo_vmware.rw_handles [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52060f3a-7260-f43b-5114-2051414cf880/disk-0.vmdk from lease info. {{(pid=62914) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 899.630057] env[62914]: DEBUG oslo_vmware.rw_handles [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52060f3a-7260-f43b-5114-2051414cf880/disk-0.vmdk for reading. {{(pid=62914) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 899.716030] env[62914]: DEBUG nova.scheduler.client.report [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.722694] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-59eed174-a6a3-47f8-8809-1d65244fff57 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.737511] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352694, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.809465] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352695, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072515} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.810010] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 899.811078] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7efd9e-c538-4ae0-82dc-56485bbda41d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.814129] env[62914]: DEBUG nova.compute.manager [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 899.837429] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] ef27919a-f205-49d9-88d8-b350fcf5cfac/ef27919a-f205-49d9-88d8-b350fcf5cfac.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.838230] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2e1d820-a449-4717-8040-f4e9a75cf610 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.857881] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Waiting for the task: (returnval){ [ 899.857881] env[62914]: value = "task-1352697" [ 899.857881] env[62914]: _type = "Task" [ 899.857881] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.865947] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352697, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.974542] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.076130] env[62914]: DEBUG oslo_vmware.api [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352696, 'name': ReconfigVM_Task, 'duration_secs': 0.248888} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.076530] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Reconfigured VM instance instance-00000047 to detach disk 2001 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 900.081786] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0c2f232-5749-4a67-8fc1-74ccfd4b2f77 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.097673] env[62914]: DEBUG oslo_vmware.api [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 900.097673] env[62914]: value = "task-1352698" [ 900.097673] env[62914]: _type = "Task" [ 900.097673] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.106210] env[62914]: DEBUG oslo_vmware.api [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352698, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.226835] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.781s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.229876] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.757s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.231389] env[62914]: INFO nova.compute.claims [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 900.240445] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352694, 'name': CreateSnapshot_Task, 'duration_secs': 0.60698} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.240779] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Created Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 900.241585] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb344789-e541-4fa4-b506-fe7a57f572be {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.253426] env[62914]: INFO nova.scheduler.client.report [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleted allocations for instance c9b94435-6546-464a-b3ba-c9e685157059 [ 900.338092] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.367454] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352697, 'name': ReconfigVM_Task, 'duration_secs': 0.418662} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.369402] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Reconfigured VM instance instance-00000051 to attach disk [datastore2] ef27919a-f205-49d9-88d8-b350fcf5cfac/ef27919a-f205-49d9-88d8-b350fcf5cfac.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.369402] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-818e4b2b-4f31-4dec-8248-c57cdc6070e9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.378110] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Waiting for the task: (returnval){ [ 900.378110] env[62914]: value = "task-1352699" [ 900.378110] env[62914]: _type = "Task" [ 900.378110] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.387472] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352699, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.611953] env[62914]: DEBUG oslo_vmware.api [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352698, 'name': ReconfigVM_Task, 'duration_secs': 0.135147} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.614337] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288222', 'volume_id': 'a3d96dcf-656d-4c56-a143-98acaae4b85f', 'name': 'volume-a3d96dcf-656d-4c56-a143-98acaae4b85f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1fe15423-f8be-4763-b55e-2cbb383ff01d', 'attached_at': '', 'detached_at': '', 'volume_id': 'a3d96dcf-656d-4c56-a143-98acaae4b85f', 'serial': 'a3d96dcf-656d-4c56-a143-98acaae4b85f'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 900.766669] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Creating linked-clone VM from snapshot {{(pid=62914) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 900.767372] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1d9abe7e-0a26-4ef7-a3a1-3849b3da341a tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "c9b94435-6546-464a-b3ba-c9e685157059" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.939s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.768639] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f6971b70-03eb-4c42-85cd-778ab54704fa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.781478] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 900.781478] env[62914]: value = "task-1352700" [ 900.781478] env[62914]: _type = "Task" [ 900.781478] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.790604] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352700, 'name': CloneVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.891029] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352699, 'name': Rename_Task, 'duration_secs': 0.27299} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.891512] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 900.891905] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5cb58fee-bfe7-455a-93b5-e904887e85af {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.899264] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Waiting for the task: (returnval){ [ 900.899264] env[62914]: value = "task-1352701" [ 900.899264] env[62914]: _type = "Task" [ 900.899264] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.912691] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352701, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.995316] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "4bf76ce0-b5d2-4184-888c-d0ef39878356" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.996131] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "4bf76ce0-b5d2-4184-888c-d0ef39878356" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.996746] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "4bf76ce0-b5d2-4184-888c-d0ef39878356-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.997141] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "4bf76ce0-b5d2-4184-888c-d0ef39878356-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.997223] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "4bf76ce0-b5d2-4184-888c-d0ef39878356-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.999821] env[62914]: INFO nova.compute.manager [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Terminating instance [ 901.002667] env[62914]: DEBUG nova.compute.manager [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 901.003081] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 901.003969] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae84e1da-1430-4c96-813e-27cc6aefa821 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.011703] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 901.011998] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a061e700-5a78-4669-92e1-ee8a5371d3e6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.018601] env[62914]: DEBUG oslo_vmware.api [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 901.018601] env[62914]: value = "task-1352702" [ 901.018601] env[62914]: _type = "Task" [ 901.018601] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.026640] env[62914]: DEBUG oslo_vmware.api [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352702, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.167977] env[62914]: DEBUG nova.objects.instance [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lazy-loading 'flavor' on Instance uuid 1fe15423-f8be-4763-b55e-2cbb383ff01d {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.291468] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352700, 'name': CloneVM_Task} progress is 94%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.409541] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352701, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.411527] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37fa691-690d-475b-bb37-2698a51cdd3f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.421043] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8222ec9a-7dbd-407c-9f52-c0c6c692a8ca {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.456379] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba93d83d-aeb6-469b-845f-4d1592afc2a3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.464626] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f98066c-cc61-4c93-9315-3d975b26f667 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.478308] env[62914]: DEBUG nova.compute.provider_tree [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.530372] env[62914]: DEBUG oslo_vmware.api [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352702, 'name': PowerOffVM_Task, 'duration_secs': 0.397033} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.531679] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 901.531978] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 901.534731] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-be078898-2228-4403-94e6-d418904a6e6d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.537327] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "f616eac4-7c75-484e-bb64-5e227a08df8c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.537616] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "f616eac4-7c75-484e-bb64-5e227a08df8c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.662260] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 901.662575] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 901.662924] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Deleting the datastore file [datastore1] 4bf76ce0-b5d2-4184-888c-d0ef39878356 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.663308] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a2cdb7a5-cec0-4ef9-95e5-4a9cfb8a8dfc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.670063] env[62914]: DEBUG oslo_vmware.api [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 901.670063] env[62914]: value = "task-1352704" [ 901.670063] env[62914]: _type = "Task" [ 901.670063] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.682625] env[62914]: DEBUG oslo_vmware.api [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352704, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.792158] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352700, 'name': CloneVM_Task} progress is 95%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.910815] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352701, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.962912] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "4bec4c88-bd1e-4cdf-9c8a-eee56088479b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.963234] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "4bec4c88-bd1e-4cdf-9c8a-eee56088479b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.981345] env[62914]: DEBUG nova.scheduler.client.report [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 902.041131] env[62914]: DEBUG nova.compute.manager [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 902.177311] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9394a2ae-e47f-4b18-a5bd-bc445d4bf601 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.261s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.182503] env[62914]: DEBUG oslo_vmware.api [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352704, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212044} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.183195] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 902.183402] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 902.183641] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 902.183853] env[62914]: INFO nova.compute.manager [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Took 1.18 seconds to destroy the instance on the hypervisor. [ 902.184124] env[62914]: DEBUG oslo.service.loopingcall [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.184328] env[62914]: DEBUG nova.compute.manager [-] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 902.184422] env[62914]: DEBUG nova.network.neutron [-] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 902.293101] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352700, 'name': CloneVM_Task, 'duration_secs': 1.190181} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.293385] env[62914]: INFO nova.virt.vmwareapi.vmops [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Created linked-clone VM from snapshot [ 902.294179] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b990b13a-703a-4405-8157-0735ae0decd1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.301795] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Uploading image 45a7cd97-03ac-4829-b7c1-2d1600121db4 {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 902.323118] env[62914]: DEBUG oslo_vmware.rw_handles [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 902.323118] env[62914]: value = "vm-288229" [ 902.323118] env[62914]: _type = "VirtualMachine" [ 902.323118] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 902.323440] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-11aece2c-5369-4988-b67d-c7847dd2b909 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.330339] env[62914]: DEBUG oslo_vmware.rw_handles [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lease: (returnval){ [ 902.330339] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52de40d6-8e72-96bb-3c7c-ec8a2dc766d2" [ 902.330339] env[62914]: _type = "HttpNfcLease" [ 902.330339] env[62914]: } obtained for exporting VM: (result){ [ 902.330339] env[62914]: value = "vm-288229" [ 902.330339] env[62914]: _type = "VirtualMachine" [ 902.330339] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 902.330664] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the lease: (returnval){ [ 902.330664] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52de40d6-8e72-96bb-3c7c-ec8a2dc766d2" [ 902.330664] env[62914]: _type = "HttpNfcLease" [ 902.330664] env[62914]: } to be ready. {{(pid=62914) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 902.336990] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 902.336990] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52de40d6-8e72-96bb-3c7c-ec8a2dc766d2" [ 902.336990] env[62914]: _type = "HttpNfcLease" [ 902.336990] env[62914]: } is initializing. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 902.397980] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.398293] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.398515] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "1fe15423-f8be-4763-b55e-2cbb383ff01d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.398700] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "1fe15423-f8be-4763-b55e-2cbb383ff01d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.398870] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "1fe15423-f8be-4763-b55e-2cbb383ff01d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.407135] env[62914]: INFO nova.compute.manager [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Terminating instance [ 902.409733] env[62914]: DEBUG nova.compute.manager [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 902.409969] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 902.412848] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d7d958-11a5-4e69-8695-0b63ab07458c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.421061] env[62914]: DEBUG oslo_vmware.api [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352701, 'name': PowerOnVM_Task, 'duration_secs': 1.270064} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.423786] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 902.423994] env[62914]: INFO nova.compute.manager [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Took 8.52 seconds to spawn the instance on the hypervisor. [ 902.424193] env[62914]: DEBUG nova.compute.manager [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 902.424493] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 902.425472] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330d3062-9d47-4e8f-9b5c-c71b6fae6885 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.428280] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d18760d-0eeb-418b-a39a-3c023c2ee043 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.431623] env[62914]: DEBUG nova.compute.manager [req-04f09e08-b858-4a73-8e06-8bdb98b58d0c req-888e5485-bb8b-44ee-a396-bb035f1af17b service nova] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Received event network-vif-deleted-189c8622-547d-4da9-94da-564df3bc382d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.431893] env[62914]: INFO nova.compute.manager [req-04f09e08-b858-4a73-8e06-8bdb98b58d0c req-888e5485-bb8b-44ee-a396-bb035f1af17b service nova] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Neutron deleted interface 189c8622-547d-4da9-94da-564df3bc382d; detaching it from the instance and deleting it from the info cache [ 902.432046] env[62914]: DEBUG nova.network.neutron [req-04f09e08-b858-4a73-8e06-8bdb98b58d0c req-888e5485-bb8b-44ee-a396-bb035f1af17b service nova] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.440582] env[62914]: DEBUG oslo_vmware.api [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 902.440582] env[62914]: value = "task-1352706" [ 902.440582] env[62914]: _type = "Task" [ 902.440582] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.449888] env[62914]: DEBUG oslo_vmware.api [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352706, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.467726] env[62914]: DEBUG nova.compute.manager [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 902.486066] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.256s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.486886] env[62914]: DEBUG nova.compute.manager [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 902.489633] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.503s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.489924] env[62914]: DEBUG nova.objects.instance [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lazy-loading 'resources' on Instance uuid 22de53a6-096b-4e8e-af91-f78588ea4c60 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.567074] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.840613] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 902.840613] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52de40d6-8e72-96bb-3c7c-ec8a2dc766d2" [ 902.840613] env[62914]: _type = "HttpNfcLease" [ 902.840613] env[62914]: } is ready. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 902.840957] env[62914]: DEBUG oslo_vmware.rw_handles [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 902.840957] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52de40d6-8e72-96bb-3c7c-ec8a2dc766d2" [ 902.840957] env[62914]: _type = "HttpNfcLease" [ 902.840957] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 902.841733] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca9291c-6881-4e2f-9ad1-f20649721523 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.850058] env[62914]: DEBUG oslo_vmware.rw_handles [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520c3647-ef9e-92b5-336d-19c1b50478df/disk-0.vmdk from lease info. {{(pid=62914) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 902.850058] env[62914]: DEBUG oslo_vmware.rw_handles [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520c3647-ef9e-92b5-336d-19c1b50478df/disk-0.vmdk for reading. {{(pid=62914) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 902.916510] env[62914]: DEBUG nova.network.neutron [-] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.936858] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5aef8e4b-0ebb-46a1-8eb2-a69210cf6a65 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.940041] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6ff043f4-0083-4408-ac7b-8d2443707cb9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.956181] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a552a95-e855-4a1c-a948-7e018a25aca1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.970675] env[62914]: INFO nova.compute.manager [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Took 13.81 seconds to build instance. [ 902.977962] env[62914]: DEBUG oslo_vmware.api [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352706, 'name': PowerOffVM_Task, 'duration_secs': 0.26453} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.980186] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 902.980379] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 902.980834] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-592e1d34-8b3c-4ba2-922e-80c1873b00bb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.994665] env[62914]: DEBUG nova.compute.utils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 902.998095] env[62914]: DEBUG nova.compute.manager [req-04f09e08-b858-4a73-8e06-8bdb98b58d0c req-888e5485-bb8b-44ee-a396-bb035f1af17b service nova] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Detach interface failed, port_id=189c8622-547d-4da9-94da-564df3bc382d, reason: Instance 4bf76ce0-b5d2-4184-888c-d0ef39878356 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 903.004131] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.004423] env[62914]: DEBUG nova.compute.manager [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 903.004608] env[62914]: DEBUG nova.network.neutron [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 903.047231] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 903.047487] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 903.047662] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Deleting the datastore file [datastore1] 1fe15423-f8be-4763-b55e-2cbb383ff01d {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.047936] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92468e7e-f52c-4d24-a4e2-660c2caf3aca {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.056947] env[62914]: DEBUG oslo_vmware.api [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 903.056947] env[62914]: value = "task-1352708" [ 903.056947] env[62914]: _type = "Task" [ 903.056947] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.060750] env[62914]: DEBUG nova.policy [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '48484c39e9ee4df8a7d5480e675f5eb8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abac39dc9aac401daf7358f128924ff4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 903.067236] env[62914]: DEBUG oslo_vmware.api [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352708, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.192298] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7890c1de-66d1-4ec6-a76f-6f3c7b970f46 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.201783] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8775b13-cbf1-4306-b826-a1997d102614 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.236379] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59224642-d71e-4cd1-ad36-4d16162889b6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.244957] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be995cab-b16a-442b-893f-e66f5dd727fc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.262980] env[62914]: DEBUG nova.compute.provider_tree [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.420957] env[62914]: INFO nova.compute.manager [-] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Took 1.24 seconds to deallocate network for instance. [ 903.436790] env[62914]: DEBUG nova.network.neutron [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Successfully created port: 0639e270-95a6-45c1-80f9-a12d293758e3 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 903.474958] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a6b5d141-3499-4626-8091-fa21382c50ad tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Lock "ef27919a-f205-49d9-88d8-b350fcf5cfac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.321s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.500042] env[62914]: DEBUG nova.compute.manager [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 903.567159] env[62914]: DEBUG oslo_vmware.api [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352708, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.34347} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.567468] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.568019] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 903.568357] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 903.568814] env[62914]: INFO nova.compute.manager [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 903.569124] env[62914]: DEBUG oslo.service.loopingcall [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.569290] env[62914]: DEBUG nova.compute.manager [-] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 903.569387] env[62914]: DEBUG nova.network.neutron [-] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 903.765850] env[62914]: DEBUG nova.scheduler.client.report [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.928027] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.259574] env[62914]: DEBUG nova.compute.manager [req-5f35f779-1e3d-4cce-b9cf-242c2ed501e6 req-154a80e9-aa07-41b2-a60c-572ab3aef1fb service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Received event network-vif-deleted-59eaafaa-3160-4236-ab11-eb6f5e3ffec8 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.259836] env[62914]: INFO nova.compute.manager [req-5f35f779-1e3d-4cce-b9cf-242c2ed501e6 req-154a80e9-aa07-41b2-a60c-572ab3aef1fb service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Neutron deleted interface 59eaafaa-3160-4236-ab11-eb6f5e3ffec8; detaching it from the instance and deleting it from the info cache [ 904.259998] env[62914]: DEBUG nova.network.neutron [req-5f35f779-1e3d-4cce-b9cf-242c2ed501e6 req-154a80e9-aa07-41b2-a60c-572ab3aef1fb service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.282172] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.792s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.284988] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.087s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.286566] env[62914]: INFO nova.compute.claims [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 904.310356] env[62914]: INFO nova.scheduler.client.report [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleted allocations for instance 22de53a6-096b-4e8e-af91-f78588ea4c60 [ 904.510196] env[62914]: DEBUG nova.compute.manager [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 904.607077] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Acquiring lock "ef27919a-f205-49d9-88d8-b350fcf5cfac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.607426] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Lock "ef27919a-f205-49d9-88d8-b350fcf5cfac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.607695] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Acquiring lock "ef27919a-f205-49d9-88d8-b350fcf5cfac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.607961] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Lock "ef27919a-f205-49d9-88d8-b350fcf5cfac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.608236] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Lock "ef27919a-f205-49d9-88d8-b350fcf5cfac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.610784] env[62914]: INFO nova.compute.manager [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Terminating instance [ 904.613422] env[62914]: DEBUG nova.compute.manager [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 904.613685] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 904.614932] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b79bfd1f-6d01-4250-aa14-3c2165515e3e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.623640] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 904.624019] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a9c55cd-5693-41f4-9adc-3a249774c071 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.630879] env[62914]: DEBUG oslo_vmware.api [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Waiting for the task: (returnval){ [ 904.630879] env[62914]: value = "task-1352709" [ 904.630879] env[62914]: _type = "Task" [ 904.630879] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.639960] env[62914]: DEBUG oslo_vmware.api [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352709, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.654027] env[62914]: DEBUG nova.network.neutron [-] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.764101] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6eabec3b-72f2-41b3-8b22-dcd4ba84a1f7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.774204] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482b676d-cbee-4e18-a6de-6f0892cc78d3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.802751] env[62914]: DEBUG nova.compute.manager [req-5f35f779-1e3d-4cce-b9cf-242c2ed501e6 req-154a80e9-aa07-41b2-a60c-572ab3aef1fb service nova] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Detach interface failed, port_id=59eaafaa-3160-4236-ab11-eb6f5e3ffec8, reason: Instance 1fe15423-f8be-4763-b55e-2cbb383ff01d could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 904.825089] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f074b01c-23d8-4b93-ac6a-7e1cd2589ef2 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "22de53a6-096b-4e8e-af91-f78588ea4c60" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.218s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.142102] env[62914]: DEBUG oslo_vmware.api [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352709, 'name': PowerOffVM_Task, 'duration_secs': 0.200455} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.142483] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 905.143087] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 905.143087] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-88f6bda7-5d9a-442d-8319-1af84877b235 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.156785] env[62914]: INFO nova.compute.manager [-] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Took 1.59 seconds to deallocate network for instance. [ 905.214438] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 905.214740] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 905.214968] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Deleting the datastore file [datastore2] ef27919a-f205-49d9-88d8-b350fcf5cfac {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 905.215305] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a807ed64-2e97-4918-8c4a-2f3328e5f720 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.222969] env[62914]: DEBUG oslo_vmware.api [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Waiting for the task: (returnval){ [ 905.222969] env[62914]: value = "task-1352711" [ 905.222969] env[62914]: _type = "Task" [ 905.222969] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.231973] env[62914]: DEBUG oslo_vmware.api [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352711, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.306120] env[62914]: DEBUG nova.network.neutron [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Successfully updated port: 0639e270-95a6-45c1-80f9-a12d293758e3 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 905.475663] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a65a3f5-6b9b-4a07-a12c-c7804ec48783 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.483086] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4fda354-078b-4133-bf45-f3e3635cf610 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.514693] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5940faef-8113-4d6d-8e8a-ed1bc7b7244a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.522478] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c53863-e347-4024-964d-240594d4d2c9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.536104] env[62914]: DEBUG nova.compute.provider_tree [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.665494] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.733995] env[62914]: DEBUG oslo_vmware.api [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Task: {'id': task-1352711, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.24974} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.734415] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 905.734678] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 905.734908] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 905.735238] env[62914]: INFO nova.compute.manager [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Took 1.12 seconds to destroy the instance on the hypervisor. [ 905.735531] env[62914]: DEBUG oslo.service.loopingcall [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.735748] env[62914]: DEBUG nova.compute.manager [-] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 905.735849] env[62914]: DEBUG nova.network.neutron [-] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 905.809019] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Acquiring lock "refresh_cache-f14f4ab4-5ef1-4f2c-ae97-ac924a446df0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.809224] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Acquired lock "refresh_cache-f14f4ab4-5ef1-4f2c-ae97-ac924a446df0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.809329] env[62914]: DEBUG nova.network.neutron [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 906.039506] env[62914]: DEBUG nova.scheduler.client.report [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 906.343087] env[62914]: DEBUG nova.network.neutron [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 906.545491] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.260s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.546145] env[62914]: DEBUG nova.compute.manager [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 906.549277] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.965s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.551406] env[62914]: INFO nova.compute.claims [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.568533] env[62914]: DEBUG nova.network.neutron [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Updating instance_info_cache with network_info: [{"id": "0639e270-95a6-45c1-80f9-a12d293758e3", "address": "fa:16:3e:c2:dd:df", "network": {"id": "294d3fca-dcb7-4243-b212-16eecd6ef9e3", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-2146544079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abac39dc9aac401daf7358f128924ff4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0639e270-95", "ovs_interfaceid": "0639e270-95a6-45c1-80f9-a12d293758e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.590056] env[62914]: DEBUG nova.network.neutron [-] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.056978] env[62914]: DEBUG nova.compute.utils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 907.062035] env[62914]: DEBUG nova.compute.manager [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 907.062035] env[62914]: DEBUG nova.network.neutron [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 907.072084] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Releasing lock "refresh_cache-f14f4ab4-5ef1-4f2c-ae97-ac924a446df0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.072962] env[62914]: DEBUG nova.compute.manager [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Instance network_info: |[{"id": "0639e270-95a6-45c1-80f9-a12d293758e3", "address": "fa:16:3e:c2:dd:df", "network": {"id": "294d3fca-dcb7-4243-b212-16eecd6ef9e3", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-2146544079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abac39dc9aac401daf7358f128924ff4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0639e270-95", "ovs_interfaceid": "0639e270-95a6-45c1-80f9-a12d293758e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 907.092813] env[62914]: INFO nova.compute.manager [-] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Took 1.36 seconds to deallocate network for instance. [ 907.103640] env[62914]: DEBUG nova.policy [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ec966b4a71a4d60b7ef3782032b3d4d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ce1d1c0e5e04202adf074388b43266f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 907.380045] env[62914]: DEBUG nova.network.neutron [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Successfully created port: 2282e6e9-7ccf-474e-9fd5-6738ed525721 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 907.565750] env[62914]: DEBUG nova.compute.manager [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 907.599864] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.676473] env[62914]: DEBUG nova.virt.hardware [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 907.676773] env[62914]: DEBUG nova.virt.hardware [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 907.677222] env[62914]: DEBUG nova.virt.hardware [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 907.677618] env[62914]: DEBUG nova.virt.hardware [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 907.677618] env[62914]: DEBUG nova.virt.hardware [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 907.677618] env[62914]: DEBUG nova.virt.hardware [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 907.677775] env[62914]: DEBUG nova.virt.hardware [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 907.678163] env[62914]: DEBUG nova.virt.hardware [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 907.678163] env[62914]: DEBUG nova.virt.hardware [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 907.678896] env[62914]: DEBUG nova.virt.hardware [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 907.678896] env[62914]: DEBUG nova.virt.hardware [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 907.681020] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5fa477c-9da2-4d12-affe-04487a206abf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.689474] env[62914]: DEBUG oslo_vmware.rw_handles [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52060f3a-7260-f43b-5114-2051414cf880/disk-0.vmdk. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 907.690443] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f876514-d36e-4749-97b7-58131ea7488b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.697764] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dacae11-9d73-48d7-96ef-3c3b5e954b9b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.705576] env[62914]: DEBUG oslo_vmware.rw_handles [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52060f3a-7260-f43b-5114-2051414cf880/disk-0.vmdk is in state: ready. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 907.705784] env[62914]: ERROR oslo_vmware.rw_handles [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52060f3a-7260-f43b-5114-2051414cf880/disk-0.vmdk due to incomplete transfer. [ 907.706755] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b7ee0be2-5fc8-4bcf-bd18-ca6f981f1f99 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.717037] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:dd:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2e0cfc48-d93b-4477-8082-69a2f7aa7701', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0639e270-95a6-45c1-80f9-a12d293758e3', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 907.724839] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Creating folder: Project (abac39dc9aac401daf7358f128924ff4). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 907.728385] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-219aa483-9efc-4ac8-bdb3-ac364d83f334 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.734650] env[62914]: DEBUG oslo_vmware.rw_handles [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52060f3a-7260-f43b-5114-2051414cf880/disk-0.vmdk. {{(pid=62914) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 907.735014] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Uploaded image 662656a7-d46d-4633-b79c-60fc2ed8c7dc to the Glance image server {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 907.737613] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Destroying the VM {{(pid=62914) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 907.737711] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a45d4afb-f202-426b-9a06-058bd247f9b6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.741071] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Created folder: Project (abac39dc9aac401daf7358f128924ff4) in parent group-v288131. [ 907.741071] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Creating folder: Instances. Parent ref: group-v288230. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 907.741939] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53e9a60e-561d-4919-b468-e885fff71e45 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.747790] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 907.747790] env[62914]: value = "task-1352713" [ 907.747790] env[62914]: _type = "Task" [ 907.747790] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.756023] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Created folder: Instances in parent group-v288230. [ 907.756023] env[62914]: DEBUG oslo.service.loopingcall [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.756023] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 907.758380] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-16bb432c-6c06-4a86-8612-5f3c17675cd6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.777687] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352713, 'name': Destroy_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.782847] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 907.782847] env[62914]: value = "task-1352715" [ 907.782847] env[62914]: _type = "Task" [ 907.782847] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.792950] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352715, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.800518] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b72758e-6410-4eee-a8be-c3c51024a261 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.807945] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beae37ed-4137-42bc-8ac3-bcc9e7cd2053 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.840063] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef571f7b-9825-477e-b9dc-f43eb119c161 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.847665] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1659c369-0b9e-4bd8-9100-a1ff86598e3e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.862073] env[62914]: DEBUG nova.compute.provider_tree [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.887536] env[62914]: DEBUG nova.compute.manager [req-c88b5226-a409-4003-8d4e-eb90f4fe885d req-28dd7292-0570-4a8a-9724-1576295d31a1 service nova] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Received event network-vif-plugged-0639e270-95a6-45c1-80f9-a12d293758e3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.887757] env[62914]: DEBUG oslo_concurrency.lockutils [req-c88b5226-a409-4003-8d4e-eb90f4fe885d req-28dd7292-0570-4a8a-9724-1576295d31a1 service nova] Acquiring lock "f14f4ab4-5ef1-4f2c-ae97-ac924a446df0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.888024] env[62914]: DEBUG oslo_concurrency.lockutils [req-c88b5226-a409-4003-8d4e-eb90f4fe885d req-28dd7292-0570-4a8a-9724-1576295d31a1 service nova] Lock "f14f4ab4-5ef1-4f2c-ae97-ac924a446df0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.888243] env[62914]: DEBUG oslo_concurrency.lockutils [req-c88b5226-a409-4003-8d4e-eb90f4fe885d req-28dd7292-0570-4a8a-9724-1576295d31a1 service nova] Lock "f14f4ab4-5ef1-4f2c-ae97-ac924a446df0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.888426] env[62914]: DEBUG nova.compute.manager [req-c88b5226-a409-4003-8d4e-eb90f4fe885d req-28dd7292-0570-4a8a-9724-1576295d31a1 service nova] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] No waiting events found dispatching network-vif-plugged-0639e270-95a6-45c1-80f9-a12d293758e3 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 907.888616] env[62914]: WARNING nova.compute.manager [req-c88b5226-a409-4003-8d4e-eb90f4fe885d req-28dd7292-0570-4a8a-9724-1576295d31a1 service nova] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Received unexpected event network-vif-plugged-0639e270-95a6-45c1-80f9-a12d293758e3 for instance with vm_state building and task_state spawning. [ 908.260464] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352713, 'name': Destroy_Task, 'duration_secs': 0.395615} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.260767] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Destroyed the VM [ 908.261027] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Deleting Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 908.261310] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1d7281f7-ae2c-4eab-8ebf-1bc6082fbdcd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.268970] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 908.268970] env[62914]: value = "task-1352716" [ 908.268970] env[62914]: _type = "Task" [ 908.268970] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.278092] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352716, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.292263] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352715, 'name': CreateVM_Task, 'duration_secs': 0.351067} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.292444] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 908.293266] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.293566] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.293922] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.294203] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94eab72d-5a72-4e8e-b80e-c338937a21d4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.299479] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Waiting for the task: (returnval){ [ 908.299479] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]522cf873-78d1-6b8c-6c2d-23d72040ecbc" [ 908.299479] env[62914]: _type = "Task" [ 908.299479] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.308798] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]522cf873-78d1-6b8c-6c2d-23d72040ecbc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.365517] env[62914]: DEBUG nova.scheduler.client.report [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 908.577360] env[62914]: DEBUG nova.compute.manager [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 908.608814] env[62914]: DEBUG nova.virt.hardware [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 908.609191] env[62914]: DEBUG nova.virt.hardware [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 908.609405] env[62914]: DEBUG nova.virt.hardware [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 908.609655] env[62914]: DEBUG nova.virt.hardware [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 908.609852] env[62914]: DEBUG nova.virt.hardware [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 908.610069] env[62914]: DEBUG nova.virt.hardware [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 908.610337] env[62914]: DEBUG nova.virt.hardware [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 908.610550] env[62914]: DEBUG nova.virt.hardware [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 908.610797] env[62914]: DEBUG nova.virt.hardware [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 908.611008] env[62914]: DEBUG nova.virt.hardware [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 908.611207] env[62914]: DEBUG nova.virt.hardware [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 908.612290] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccff712e-bfd1-47b8-a4ce-77a51a8dc51b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.621293] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ac86b1-aa5e-40c4-9583-16401a858d18 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.779403] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352716, 'name': RemoveSnapshot_Task, 'duration_secs': 0.327223} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.780576] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Deleted Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 908.780576] env[62914]: DEBUG nova.compute.manager [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.780935] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9075c4c7-608e-4f4f-819c-4095a6295bc5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.811065] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]522cf873-78d1-6b8c-6c2d-23d72040ecbc, 'name': SearchDatastore_Task, 'duration_secs': 0.011081} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.812023] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.812023] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 908.812023] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.812223] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.812223] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 908.812499] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-93158b1b-e5b9-4ed5-b598-ab490775efe0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.821776] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 908.822409] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 908.823748] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74a30e67-a14c-4d91-bc87-6f8490f04832 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.828744] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Waiting for the task: (returnval){ [ 908.828744] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d3bb35-6aeb-c751-2790-c25c1acc81a3" [ 908.828744] env[62914]: _type = "Task" [ 908.828744] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.837379] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d3bb35-6aeb-c751-2790-c25c1acc81a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.871128] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.871709] env[62914]: DEBUG nova.compute.manager [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 908.874982] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.900s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.875858] env[62914]: INFO nova.compute.claims [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.936371] env[62914]: DEBUG nova.network.neutron [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Successfully updated port: 2282e6e9-7ccf-474e-9fd5-6738ed525721 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 909.293965] env[62914]: INFO nova.compute.manager [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Shelve offloading [ 909.295568] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 909.295817] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17da4480-ee54-4a35-81a4-b416438f3b17 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.303652] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 909.303652] env[62914]: value = "task-1352717" [ 909.303652] env[62914]: _type = "Task" [ 909.303652] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.313770] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] VM already powered off {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 909.313950] env[62914]: DEBUG nova.compute.manager [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.314687] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f766f2f-aa8a-4b4a-b6d8-99b48ad1c01a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.320325] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.320487] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquired lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.320657] env[62914]: DEBUG nova.network.neutron [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 909.338373] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d3bb35-6aeb-c751-2790-c25c1acc81a3, 'name': SearchDatastore_Task, 'duration_secs': 0.00927} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.339158] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc7b5602-6d94-4342-af53-c10b39f9b531 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.344439] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Waiting for the task: (returnval){ [ 909.344439] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f4d5f0-33cb-85c4-cf70-ef1b0b746ca9" [ 909.344439] env[62914]: _type = "Task" [ 909.344439] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.352624] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f4d5f0-33cb-85c4-cf70-ef1b0b746ca9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.381646] env[62914]: DEBUG nova.compute.utils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.384919] env[62914]: DEBUG nova.compute.manager [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 909.385104] env[62914]: DEBUG nova.network.neutron [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 909.425760] env[62914]: DEBUG nova.policy [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ec966b4a71a4d60b7ef3782032b3d4d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ce1d1c0e5e04202adf074388b43266f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 909.439395] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "refresh_cache-ab036e1f-a62a-490a-818c-2e1256c781f6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.439558] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquired lock "refresh_cache-ab036e1f-a62a-490a-818c-2e1256c781f6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.439674] env[62914]: DEBUG nova.network.neutron [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 909.498483] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "13954007-b5ee-4254-b99e-5fbb2f6e0add" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.498769] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "13954007-b5ee-4254-b99e-5fbb2f6e0add" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.679215] env[62914]: DEBUG oslo_vmware.rw_handles [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520c3647-ef9e-92b5-336d-19c1b50478df/disk-0.vmdk. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 909.682855] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b5ea70-f11e-4376-b010-aebd9331848e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.690799] env[62914]: DEBUG oslo_vmware.rw_handles [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520c3647-ef9e-92b5-336d-19c1b50478df/disk-0.vmdk is in state: ready. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 909.691032] env[62914]: ERROR oslo_vmware.rw_handles [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520c3647-ef9e-92b5-336d-19c1b50478df/disk-0.vmdk due to incomplete transfer. [ 909.691253] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-aea827f7-3bec-4495-90be-116e9bb01289 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.698311] env[62914]: DEBUG oslo_vmware.rw_handles [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520c3647-ef9e-92b5-336d-19c1b50478df/disk-0.vmdk. {{(pid=62914) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 909.698862] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Uploaded image 45a7cd97-03ac-4829-b7c1-2d1600121db4 to the Glance image server {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 909.700445] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Destroying the VM {{(pid=62914) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 909.700730] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-aabb85bf-3905-44c9-8f46-c9461f46d10b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.704102] env[62914]: DEBUG nova.network.neutron [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Successfully created port: ad5c411a-d657-4b8e-8cb7-1ef6d90a0bc2 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 909.707546] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 909.707546] env[62914]: value = "task-1352718" [ 909.707546] env[62914]: _type = "Task" [ 909.707546] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.715419] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352718, 'name': Destroy_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.865816] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f4d5f0-33cb-85c4-cf70-ef1b0b746ca9, 'name': SearchDatastore_Task, 'duration_secs': 0.009792} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.868507] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.869494] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] f14f4ab4-5ef1-4f2c-ae97-ac924a446df0/f14f4ab4-5ef1-4f2c-ae97-ac924a446df0.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 909.869494] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3cd42f1-8b79-4f52-9522-aaac6ff67312 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.878960] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Waiting for the task: (returnval){ [ 909.878960] env[62914]: value = "task-1352719" [ 909.878960] env[62914]: _type = "Task" [ 909.878960] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.889604] env[62914]: DEBUG nova.compute.manager [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 909.895727] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352719, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.969108] env[62914]: DEBUG nova.compute.manager [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Received event network-changed-0639e270-95a6-45c1-80f9-a12d293758e3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.969159] env[62914]: DEBUG nova.compute.manager [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Refreshing instance network info cache due to event network-changed-0639e270-95a6-45c1-80f9-a12d293758e3. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 909.969353] env[62914]: DEBUG oslo_concurrency.lockutils [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] Acquiring lock "refresh_cache-f14f4ab4-5ef1-4f2c-ae97-ac924a446df0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.969508] env[62914]: DEBUG oslo_concurrency.lockutils [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] Acquired lock "refresh_cache-f14f4ab4-5ef1-4f2c-ae97-ac924a446df0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.969748] env[62914]: DEBUG nova.network.neutron [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Refreshing network info cache for port 0639e270-95a6-45c1-80f9-a12d293758e3 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 910.000912] env[62914]: DEBUG nova.compute.manager [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 910.019616] env[62914]: DEBUG nova.network.neutron [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 910.153700] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77624556-4cb0-41c9-8e7e-83bedbedb55f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.164247] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d42aec7-3836-429b-9f13-452e5c55e010 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.219828] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d6dde9-960e-4b81-8bae-f667ede8305a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.233842] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f68b8a19-a572-4bd4-a2e9-aec3bb06de88 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.237892] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352718, 'name': Destroy_Task, 'duration_secs': 0.441162} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.238204] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Destroyed the VM [ 910.238448] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Deleting Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 910.239150] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5360f1e8-39ba-4751-96e2-d02d04a7803f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.249162] env[62914]: DEBUG nova.compute.provider_tree [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.263015] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 910.263015] env[62914]: value = "task-1352720" [ 910.263015] env[62914]: _type = "Task" [ 910.263015] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.270507] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352720, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.335222] env[62914]: DEBUG nova.network.neutron [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updating instance_info_cache with network_info: [{"id": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "address": "fa:16:3e:91:13:35", "network": {"id": "c4298e11-ab16-4b37-be29-48fc406e74b0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-695272838-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.236", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "15315f8e77d749a4b2b0211c3b22b546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe38bb7e-8bcb-419d-868f-0dc105c69651", "external-id": "nsx-vlan-transportzone-432", "segmentation_id": 432, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57172af7-5e", "ovs_interfaceid": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.366192] env[62914]: DEBUG nova.network.neutron [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Updating instance_info_cache with network_info: [{"id": "2282e6e9-7ccf-474e-9fd5-6738ed525721", "address": "fa:16:3e:ec:98:9f", "network": {"id": "f59afc5a-2dbe-4ddd-962e-6ec7eb665075", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-240973820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ce1d1c0e5e04202adf074388b43266f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "305ccd93-08cb-4658-845c-d9b64952daf7", "external-id": "nsx-vlan-transportzone-490", "segmentation_id": 490, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2282e6e9-7c", "ovs_interfaceid": "2282e6e9-7ccf-474e-9fd5-6738ed525721", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.392013] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352719, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.521100] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.560579] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "aec39e18-7796-4be3-af74-478df3a78f8f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.560579] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "aec39e18-7796-4be3-af74-478df3a78f8f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.560579] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "aec39e18-7796-4be3-af74-478df3a78f8f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.560579] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "aec39e18-7796-4be3-af74-478df3a78f8f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.560579] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "aec39e18-7796-4be3-af74-478df3a78f8f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.561838] env[62914]: INFO nova.compute.manager [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Terminating instance [ 910.569019] env[62914]: DEBUG nova.compute.manager [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 910.569019] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 910.569019] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e30b04-f9b7-4121-b581-fcb76d4bf145 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.581786] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 910.581786] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a51c6948-36d2-40fa-99ee-36863819799c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.712731] env[62914]: DEBUG nova.network.neutron [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Updated VIF entry in instance network info cache for port 0639e270-95a6-45c1-80f9-a12d293758e3. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 910.713190] env[62914]: DEBUG nova.network.neutron [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Updating instance_info_cache with network_info: [{"id": "0639e270-95a6-45c1-80f9-a12d293758e3", "address": "fa:16:3e:c2:dd:df", "network": {"id": "294d3fca-dcb7-4243-b212-16eecd6ef9e3", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-2146544079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abac39dc9aac401daf7358f128924ff4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0639e270-95", "ovs_interfaceid": "0639e270-95a6-45c1-80f9-a12d293758e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.752343] env[62914]: DEBUG nova.scheduler.client.report [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.770603] env[62914]: DEBUG oslo_vmware.api [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352720, 'name': RemoveSnapshot_Task, 'duration_secs': 0.473561} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.771526] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Deleted Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 910.771809] env[62914]: INFO nova.compute.manager [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Took 12.11 seconds to snapshot the instance on the hypervisor. [ 910.838511] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Releasing lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.869445] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Releasing lock "refresh_cache-ab036e1f-a62a-490a-818c-2e1256c781f6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.869956] env[62914]: DEBUG nova.compute.manager [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Instance network_info: |[{"id": "2282e6e9-7ccf-474e-9fd5-6738ed525721", "address": "fa:16:3e:ec:98:9f", "network": {"id": "f59afc5a-2dbe-4ddd-962e-6ec7eb665075", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-240973820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ce1d1c0e5e04202adf074388b43266f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "305ccd93-08cb-4658-845c-d9b64952daf7", "external-id": "nsx-vlan-transportzone-490", "segmentation_id": 490, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2282e6e9-7c", "ovs_interfaceid": "2282e6e9-7ccf-474e-9fd5-6738ed525721", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 910.873673] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:98:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '305ccd93-08cb-4658-845c-d9b64952daf7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2282e6e9-7ccf-474e-9fd5-6738ed525721', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 910.881250] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Creating folder: Project (2ce1d1c0e5e04202adf074388b43266f). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 910.882627] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-595d5cba-c8f4-4d2f-b7b6-925374110afb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.884584] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 910.884778] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 910.884968] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleting the datastore file [datastore1] aec39e18-7796-4be3-af74-478df3a78f8f {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.888075] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e96bc25f-2a6e-45b3-8848-f36ee90036c0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.894944] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352719, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.596962} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.896121] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] f14f4ab4-5ef1-4f2c-ae97-ac924a446df0/f14f4ab4-5ef1-4f2c-ae97-ac924a446df0.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 910.896345] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 910.896633] env[62914]: DEBUG oslo_vmware.api [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 910.896633] env[62914]: value = "task-1352723" [ 910.896633] env[62914]: _type = "Task" [ 910.896633] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.896812] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6aaff695-b934-44b9-a8db-842d04ef6812 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.902549] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Created folder: Project (2ce1d1c0e5e04202adf074388b43266f) in parent group-v288131. [ 910.902768] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Creating folder: Instances. Parent ref: group-v288233. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 910.904245] env[62914]: DEBUG nova.compute.manager [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 910.906066] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d54193f-8d3e-4cb2-88f3-98ac6d4deac6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.908907] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Waiting for the task: (returnval){ [ 910.908907] env[62914]: value = "task-1352724" [ 910.908907] env[62914]: _type = "Task" [ 910.908907] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.916300] env[62914]: DEBUG oslo_vmware.api [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352723, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.917786] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Created folder: Instances in parent group-v288233. [ 910.918038] env[62914]: DEBUG oslo.service.loopingcall [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.920136] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 910.922899] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b0ce03aa-4652-43e4-b867-7ff465c601d9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.937471] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352724, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.946112] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 910.946112] env[62914]: value = "task-1352726" [ 910.946112] env[62914]: _type = "Task" [ 910.946112] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.948957] env[62914]: DEBUG nova.virt.hardware [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.948957] env[62914]: DEBUG nova.virt.hardware [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.949532] env[62914]: DEBUG nova.virt.hardware [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.949532] env[62914]: DEBUG nova.virt.hardware [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.949532] env[62914]: DEBUG nova.virt.hardware [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.949701] env[62914]: DEBUG nova.virt.hardware [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.949863] env[62914]: DEBUG nova.virt.hardware [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.950054] env[62914]: DEBUG nova.virt.hardware [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.950259] env[62914]: DEBUG nova.virt.hardware [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.950455] env[62914]: DEBUG nova.virt.hardware [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.950662] env[62914]: DEBUG nova.virt.hardware [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.951611] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7cb408a-7d21-447b-9973-1e221effe556 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.962463] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352726, 'name': CreateVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.966142] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759d2ffb-ecee-4724-bb8d-9c998f02470f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.101829] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "412c706a-31ce-4a33-914d-5762b02b4c63" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.102125] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "412c706a-31ce-4a33-914d-5762b02b4c63" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.122993] env[62914]: DEBUG nova.compute.manager [req-20aff97c-2b7a-4c1b-837c-d5d62e417b98 req-ad5346e9-0975-4cdd-a184-094f3d86630c service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Received event network-vif-unplugged-57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.123119] env[62914]: DEBUG oslo_concurrency.lockutils [req-20aff97c-2b7a-4c1b-837c-d5d62e417b98 req-ad5346e9-0975-4cdd-a184-094f3d86630c service nova] Acquiring lock "c7cf71d7-b823-471f-949f-9ef2b539d972-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.123329] env[62914]: DEBUG oslo_concurrency.lockutils [req-20aff97c-2b7a-4c1b-837c-d5d62e417b98 req-ad5346e9-0975-4cdd-a184-094f3d86630c service nova] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.123512] env[62914]: DEBUG oslo_concurrency.lockutils [req-20aff97c-2b7a-4c1b-837c-d5d62e417b98 req-ad5346e9-0975-4cdd-a184-094f3d86630c service nova] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.123680] env[62914]: DEBUG nova.compute.manager [req-20aff97c-2b7a-4c1b-837c-d5d62e417b98 req-ad5346e9-0975-4cdd-a184-094f3d86630c service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] No waiting events found dispatching network-vif-unplugged-57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.123861] env[62914]: WARNING nova.compute.manager [req-20aff97c-2b7a-4c1b-837c-d5d62e417b98 req-ad5346e9-0975-4cdd-a184-094f3d86630c service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Received unexpected event network-vif-unplugged-57172af7-5ed6-46d9-9d7b-3ef100c530f7 for instance with vm_state shelved and task_state shelving_offloading. [ 911.165948] env[62914]: DEBUG nova.compute.manager [req-792bdf08-6db0-4021-841f-cef930450701 req-22a23659-dee3-478d-ac5a-5be3b8b0eadb service nova] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Received event network-vif-plugged-ad5c411a-d657-4b8e-8cb7-1ef6d90a0bc2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.165948] env[62914]: DEBUG oslo_concurrency.lockutils [req-792bdf08-6db0-4021-841f-cef930450701 req-22a23659-dee3-478d-ac5a-5be3b8b0eadb service nova] Acquiring lock "b8aebca4-2724-419e-aaf7-d180e299a459-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.165948] env[62914]: DEBUG oslo_concurrency.lockutils [req-792bdf08-6db0-4021-841f-cef930450701 req-22a23659-dee3-478d-ac5a-5be3b8b0eadb service nova] Lock "b8aebca4-2724-419e-aaf7-d180e299a459-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.165948] env[62914]: DEBUG oslo_concurrency.lockutils [req-792bdf08-6db0-4021-841f-cef930450701 req-22a23659-dee3-478d-ac5a-5be3b8b0eadb service nova] Lock "b8aebca4-2724-419e-aaf7-d180e299a459-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.166316] env[62914]: DEBUG nova.compute.manager [req-792bdf08-6db0-4021-841f-cef930450701 req-22a23659-dee3-478d-ac5a-5be3b8b0eadb service nova] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] No waiting events found dispatching network-vif-plugged-ad5c411a-d657-4b8e-8cb7-1ef6d90a0bc2 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.166627] env[62914]: WARNING nova.compute.manager [req-792bdf08-6db0-4021-841f-cef930450701 req-22a23659-dee3-478d-ac5a-5be3b8b0eadb service nova] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Received unexpected event network-vif-plugged-ad5c411a-d657-4b8e-8cb7-1ef6d90a0bc2 for instance with vm_state building and task_state spawning. [ 911.217492] env[62914]: DEBUG oslo_concurrency.lockutils [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] Releasing lock "refresh_cache-f14f4ab4-5ef1-4f2c-ae97-ac924a446df0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.217492] env[62914]: DEBUG nova.compute.manager [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Received event network-vif-deleted-52773cfb-b19a-47c3-aaae-7df97e237662 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.217492] env[62914]: DEBUG nova.compute.manager [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Received event network-vif-plugged-2282e6e9-7ccf-474e-9fd5-6738ed525721 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.217492] env[62914]: DEBUG oslo_concurrency.lockutils [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] Acquiring lock "ab036e1f-a62a-490a-818c-2e1256c781f6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.217492] env[62914]: DEBUG oslo_concurrency.lockutils [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] Lock "ab036e1f-a62a-490a-818c-2e1256c781f6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.217492] env[62914]: DEBUG oslo_concurrency.lockutils [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] Lock "ab036e1f-a62a-490a-818c-2e1256c781f6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.217492] env[62914]: DEBUG nova.compute.manager [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] No waiting events found dispatching network-vif-plugged-2282e6e9-7ccf-474e-9fd5-6738ed525721 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.217722] env[62914]: WARNING nova.compute.manager [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Received unexpected event network-vif-plugged-2282e6e9-7ccf-474e-9fd5-6738ed525721 for instance with vm_state building and task_state spawning. [ 911.217937] env[62914]: DEBUG nova.compute.manager [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Received event network-changed-2282e6e9-7ccf-474e-9fd5-6738ed525721 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.218179] env[62914]: DEBUG nova.compute.manager [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Refreshing instance network info cache due to event network-changed-2282e6e9-7ccf-474e-9fd5-6738ed525721. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 911.218428] env[62914]: DEBUG oslo_concurrency.lockutils [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] Acquiring lock "refresh_cache-ab036e1f-a62a-490a-818c-2e1256c781f6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.218607] env[62914]: DEBUG oslo_concurrency.lockutils [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] Acquired lock "refresh_cache-ab036e1f-a62a-490a-818c-2e1256c781f6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.218847] env[62914]: DEBUG nova.network.neutron [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Refreshing network info cache for port 2282e6e9-7ccf-474e-9fd5-6738ed525721 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 911.221257] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 911.222364] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f487878-9132-4805-a87e-01656ceeb63a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.230851] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 911.231206] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-15f39655-a3b2-4355-9712-353989350249 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.257251] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.383s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.257956] env[62914]: DEBUG nova.compute.manager [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 911.260953] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.923s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.262504] env[62914]: INFO nova.compute.claims [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 911.275941] env[62914]: DEBUG nova.compute.manager [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Instance disappeared during snapshot {{(pid=62914) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 911.286636] env[62914]: DEBUG nova.compute.manager [None req-ffe4f8d0-8f38-4892-a070-f4bfcaecbf86 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Image not found during clean up 45a7cd97-03ac-4829-b7c1-2d1600121db4 {{(pid=62914) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 911.313426] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 911.313705] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 911.313897] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Deleting the datastore file [datastore1] c7cf71d7-b823-471f-949f-9ef2b539d972 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 911.314185] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4bd4fd1-f7f2-4f6b-91ff-44268ec7a790 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.320562] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 911.320562] env[62914]: value = "task-1352728" [ 911.320562] env[62914]: _type = "Task" [ 911.320562] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.328321] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352728, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.408017] env[62914]: DEBUG oslo_vmware.api [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352723, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.189209} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.408359] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.408561] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 911.408870] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 911.409054] env[62914]: INFO nova.compute.manager [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Took 0.84 seconds to destroy the instance on the hypervisor. [ 911.409399] env[62914]: DEBUG oslo.service.loopingcall [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.409672] env[62914]: DEBUG nova.compute.manager [-] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.409804] env[62914]: DEBUG nova.network.neutron [-] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 911.420790] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352724, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070995} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.421059] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.421784] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e916114a-b30e-4643-8d63-d6fc759e1654 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.444704] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] f14f4ab4-5ef1-4f2c-ae97-ac924a446df0/f14f4ab4-5ef1-4f2c-ae97-ac924a446df0.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.445013] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d48c308d-d117-45cc-b7c3-01bd8dfe4727 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.467046] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352726, 'name': CreateVM_Task, 'duration_secs': 0.369909} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.468076] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 911.468403] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Waiting for the task: (returnval){ [ 911.468403] env[62914]: value = "task-1352729" [ 911.468403] env[62914]: _type = "Task" [ 911.468403] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.469011] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.469192] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.469501] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 911.469781] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e4c534f-c556-4157-8bba-a7523ba3da6c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.476767] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 911.476767] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52434b35-423b-6ae1-0f7a-acd1fd58d961" [ 911.476767] env[62914]: _type = "Task" [ 911.476767] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.479554] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352729, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.489569] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52434b35-423b-6ae1-0f7a-acd1fd58d961, 'name': SearchDatastore_Task, 'duration_secs': 0.008932} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.489569] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.489569] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.489810] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.489960] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.490172] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.490418] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8c9a312-b5e7-496e-8ddb-fe4d3535af58 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.496977] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.497188] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 911.497878] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0ceb050-a87d-4a17-9d79-cac6e9761329 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.502259] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 911.502259] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52184e7a-241f-c7e6-e7c2-3e09daaf98c7" [ 911.502259] env[62914]: _type = "Task" [ 911.502259] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.509448] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52184e7a-241f-c7e6-e7c2-3e09daaf98c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.604738] env[62914]: DEBUG nova.compute.manager [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 911.717566] env[62914]: DEBUG nova.network.neutron [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Successfully updated port: ad5c411a-d657-4b8e-8cb7-1ef6d90a0bc2 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 911.767366] env[62914]: DEBUG nova.compute.utils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 911.772446] env[62914]: DEBUG nova.compute.manager [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 911.772446] env[62914]: DEBUG nova.network.neutron [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 911.810089] env[62914]: DEBUG nova.policy [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ec966b4a71a4d60b7ef3782032b3d4d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ce1d1c0e5e04202adf074388b43266f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 911.832432] env[62914]: DEBUG oslo_vmware.api [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352728, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124991} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.832775] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.833081] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 911.833312] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 911.859750] env[62914]: INFO nova.scheduler.client.report [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Deleted allocations for instance c7cf71d7-b823-471f-949f-9ef2b539d972 [ 911.930177] env[62914]: DEBUG nova.network.neutron [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Updated VIF entry in instance network info cache for port 2282e6e9-7ccf-474e-9fd5-6738ed525721. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 911.930541] env[62914]: DEBUG nova.network.neutron [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Updating instance_info_cache with network_info: [{"id": "2282e6e9-7ccf-474e-9fd5-6738ed525721", "address": "fa:16:3e:ec:98:9f", "network": {"id": "f59afc5a-2dbe-4ddd-962e-6ec7eb665075", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-240973820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ce1d1c0e5e04202adf074388b43266f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "305ccd93-08cb-4658-845c-d9b64952daf7", "external-id": "nsx-vlan-transportzone-490", "segmentation_id": 490, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2282e6e9-7c", "ovs_interfaceid": "2282e6e9-7ccf-474e-9fd5-6738ed525721", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.979586] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352729, 'name': ReconfigVM_Task, 'duration_secs': 0.294607} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.979880] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Reconfigured VM instance instance-00000052 to attach disk [datastore2] f14f4ab4-5ef1-4f2c-ae97-ac924a446df0/f14f4ab4-5ef1-4f2c-ae97-ac924a446df0.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 911.980545] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-779b74d1-c395-43f3-8a33-b79cadf5e508 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.986892] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Waiting for the task: (returnval){ [ 911.986892] env[62914]: value = "task-1352730" [ 911.986892] env[62914]: _type = "Task" [ 911.986892] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.993124] env[62914]: DEBUG nova.compute.manager [req-fdf4defb-587e-4c3d-8a78-e9ad13de6436 req-0e7f7f2c-adcb-4806-a861-013316d8b538 service nova] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Received event network-changed-ad5c411a-d657-4b8e-8cb7-1ef6d90a0bc2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.993318] env[62914]: DEBUG nova.compute.manager [req-fdf4defb-587e-4c3d-8a78-e9ad13de6436 req-0e7f7f2c-adcb-4806-a861-013316d8b538 service nova] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Refreshing instance network info cache due to event network-changed-ad5c411a-d657-4b8e-8cb7-1ef6d90a0bc2. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 911.993598] env[62914]: DEBUG oslo_concurrency.lockutils [req-fdf4defb-587e-4c3d-8a78-e9ad13de6436 req-0e7f7f2c-adcb-4806-a861-013316d8b538 service nova] Acquiring lock "refresh_cache-b8aebca4-2724-419e-aaf7-d180e299a459" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.994320] env[62914]: DEBUG oslo_concurrency.lockutils [req-fdf4defb-587e-4c3d-8a78-e9ad13de6436 req-0e7f7f2c-adcb-4806-a861-013316d8b538 service nova] Acquired lock "refresh_cache-b8aebca4-2724-419e-aaf7-d180e299a459" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.994320] env[62914]: DEBUG nova.network.neutron [req-fdf4defb-587e-4c3d-8a78-e9ad13de6436 req-0e7f7f2c-adcb-4806-a861-013316d8b538 service nova] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Refreshing network info cache for port ad5c411a-d657-4b8e-8cb7-1ef6d90a0bc2 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 912.000167] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352730, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.012018] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52184e7a-241f-c7e6-e7c2-3e09daaf98c7, 'name': SearchDatastore_Task, 'duration_secs': 0.008649} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.012795] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24107964-b292-40d4-8cbe-0764dddf1782 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.018606] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 912.018606] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52369312-10ce-4328-f019-5a2be0ae34f8" [ 912.018606] env[62914]: _type = "Task" [ 912.018606] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.026301] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52369312-10ce-4328-f019-5a2be0ae34f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.044297] env[62914]: DEBUG nova.network.neutron [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Successfully created port: d621ec78-18bc-443c-bc86-a1bd2846b28d {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 912.124664] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.208450] env[62914]: DEBUG nova.network.neutron [-] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.219940] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "refresh_cache-b8aebca4-2724-419e-aaf7-d180e299a459" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.277418] env[62914]: DEBUG nova.compute.manager [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 912.365795] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.433787] env[62914]: DEBUG oslo_concurrency.lockutils [req-c84a70e2-efb9-4fe1-bc74-6c56543e8a8e req-64d3f8a4-88d7-4a30-bec2-3e170464a36b service nova] Releasing lock "refresh_cache-ab036e1f-a62a-490a-818c-2e1256c781f6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.438458] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dec9b85-b18d-49ab-8505-569dd1e6fafe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.445857] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67da34bb-a02b-4c7f-8c7b-637ba91b5c7f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.475877] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14c7794-f614-41b7-8a9f-7152f07e5c7a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.483140] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac67b680-40a0-49de-b14d-7ee3084c232a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.501454] env[62914]: DEBUG nova.compute.provider_tree [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.507275] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352730, 'name': Rename_Task, 'duration_secs': 0.130712} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.507503] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 912.507670] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab220251-ebde-494f-9507-c8945d23d8f3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.513939] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Waiting for the task: (returnval){ [ 912.513939] env[62914]: value = "task-1352731" [ 912.513939] env[62914]: _type = "Task" [ 912.513939] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.521347] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352731, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.529156] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52369312-10ce-4328-f019-5a2be0ae34f8, 'name': SearchDatastore_Task, 'duration_secs': 0.008259} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.529402] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.529653] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] ab036e1f-a62a-490a-818c-2e1256c781f6/ab036e1f-a62a-490a-818c-2e1256c781f6.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 912.530424] env[62914]: DEBUG nova.network.neutron [req-fdf4defb-587e-4c3d-8a78-e9ad13de6436 req-0e7f7f2c-adcb-4806-a861-013316d8b538 service nova] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 912.531990] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88e8a486-61ff-4aff-9f77-f00c47a76338 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.538185] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 912.538185] env[62914]: value = "task-1352732" [ 912.538185] env[62914]: _type = "Task" [ 912.538185] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.546115] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352732, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.606019] env[62914]: DEBUG nova.network.neutron [req-fdf4defb-587e-4c3d-8a78-e9ad13de6436 req-0e7f7f2c-adcb-4806-a861-013316d8b538 service nova] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.711864] env[62914]: INFO nova.compute.manager [-] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Took 1.30 seconds to deallocate network for instance. [ 913.005374] env[62914]: DEBUG nova.scheduler.client.report [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 913.026879] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352731, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.051128] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352732, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.109215] env[62914]: DEBUG oslo_concurrency.lockutils [req-fdf4defb-587e-4c3d-8a78-e9ad13de6436 req-0e7f7f2c-adcb-4806-a861-013316d8b538 service nova] Releasing lock "refresh_cache-b8aebca4-2724-419e-aaf7-d180e299a459" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.109645] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquired lock "refresh_cache-b8aebca4-2724-419e-aaf7-d180e299a459" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.109815] env[62914]: DEBUG nova.network.neutron [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 913.151853] env[62914]: DEBUG nova.compute.manager [req-565bf72d-5974-46ed-8c3f-08c49abf0ecc req-1a576fed-04f2-4ecf-9aed-9c49765cb15e service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Received event network-changed-57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.152095] env[62914]: DEBUG nova.compute.manager [req-565bf72d-5974-46ed-8c3f-08c49abf0ecc req-1a576fed-04f2-4ecf-9aed-9c49765cb15e service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Refreshing instance network info cache due to event network-changed-57172af7-5ed6-46d9-9d7b-3ef100c530f7. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 913.152315] env[62914]: DEBUG oslo_concurrency.lockutils [req-565bf72d-5974-46ed-8c3f-08c49abf0ecc req-1a576fed-04f2-4ecf-9aed-9c49765cb15e service nova] Acquiring lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.152462] env[62914]: DEBUG oslo_concurrency.lockutils [req-565bf72d-5974-46ed-8c3f-08c49abf0ecc req-1a576fed-04f2-4ecf-9aed-9c49765cb15e service nova] Acquired lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.152624] env[62914]: DEBUG nova.network.neutron [req-565bf72d-5974-46ed-8c3f-08c49abf0ecc req-1a576fed-04f2-4ecf-9aed-9c49765cb15e service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Refreshing network info cache for port 57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 913.218991] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.286421] env[62914]: DEBUG nova.compute.manager [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 913.310894] env[62914]: DEBUG nova.virt.hardware [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 913.311202] env[62914]: DEBUG nova.virt.hardware [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 913.311364] env[62914]: DEBUG nova.virt.hardware [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 913.311547] env[62914]: DEBUG nova.virt.hardware [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 913.311695] env[62914]: DEBUG nova.virt.hardware [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 913.311922] env[62914]: DEBUG nova.virt.hardware [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 913.312170] env[62914]: DEBUG nova.virt.hardware [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 913.312338] env[62914]: DEBUG nova.virt.hardware [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 913.312494] env[62914]: DEBUG nova.virt.hardware [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 913.312657] env[62914]: DEBUG nova.virt.hardware [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 913.312830] env[62914]: DEBUG nova.virt.hardware [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 913.313705] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045c51e3-d725-4a36-8c36-86be4841e8cc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.321544] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-063e5f57-cef1-4710-a76c-aa5efb97df45 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.509227] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.248s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.509748] env[62914]: DEBUG nova.compute.manager [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 913.513065] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.946s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.514312] env[62914]: INFO nova.compute.claims [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.525249] env[62914]: DEBUG oslo_vmware.api [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352731, 'name': PowerOnVM_Task, 'duration_secs': 0.767407} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.527710] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 913.527710] env[62914]: INFO nova.compute.manager [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Took 9.02 seconds to spawn the instance on the hypervisor. [ 913.527710] env[62914]: DEBUG nova.compute.manager [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 913.527710] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14dd4a11-1b9d-4089-8cee-37cc741fd6ea {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.548608] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352732, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.563642} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.548608] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] ab036e1f-a62a-490a-818c-2e1256c781f6/ab036e1f-a62a-490a-818c-2e1256c781f6.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 913.548608] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 913.548608] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3c84f92a-1578-4034-8fd1-2fe34be25010 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.553329] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 913.553329] env[62914]: value = "task-1352733" [ 913.553329] env[62914]: _type = "Task" [ 913.553329] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.560938] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352733, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.561623] env[62914]: DEBUG nova.network.neutron [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Successfully updated port: d621ec78-18bc-443c-bc86-a1bd2846b28d {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 913.645976] env[62914]: DEBUG nova.network.neutron [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 913.809691] env[62914]: DEBUG nova.network.neutron [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Updating instance_info_cache with network_info: [{"id": "ad5c411a-d657-4b8e-8cb7-1ef6d90a0bc2", "address": "fa:16:3e:a1:36:fa", "network": {"id": "f59afc5a-2dbe-4ddd-962e-6ec7eb665075", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-240973820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ce1d1c0e5e04202adf074388b43266f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "305ccd93-08cb-4658-845c-d9b64952daf7", "external-id": "nsx-vlan-transportzone-490", "segmentation_id": 490, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad5c411a-d6", "ovs_interfaceid": "ad5c411a-d657-4b8e-8cb7-1ef6d90a0bc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.911093] env[62914]: DEBUG nova.network.neutron [req-565bf72d-5974-46ed-8c3f-08c49abf0ecc req-1a576fed-04f2-4ecf-9aed-9c49765cb15e service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updated VIF entry in instance network info cache for port 57172af7-5ed6-46d9-9d7b-3ef100c530f7. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 913.911488] env[62914]: DEBUG nova.network.neutron [req-565bf72d-5974-46ed-8c3f-08c49abf0ecc req-1a576fed-04f2-4ecf-9aed-9c49765cb15e service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updating instance_info_cache with network_info: [{"id": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "address": "fa:16:3e:91:13:35", "network": {"id": "c4298e11-ab16-4b37-be29-48fc406e74b0", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-695272838-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.236", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "15315f8e77d749a4b2b0211c3b22b546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap57172af7-5e", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.018555] env[62914]: DEBUG nova.compute.utils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 914.020522] env[62914]: DEBUG nova.compute.manager [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 914.020896] env[62914]: DEBUG nova.network.neutron [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 914.040931] env[62914]: INFO nova.compute.manager [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Took 16.59 seconds to build instance. [ 914.063040] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352733, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06292} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.063380] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "refresh_cache-1f7b1609-08e0-4439-aea0-b4a2ec808780" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.063514] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquired lock "refresh_cache-1f7b1609-08e0-4439-aea0-b4a2ec808780" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.063655] env[62914]: DEBUG nova.network.neutron [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.064691] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 914.065538] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d69036d7-cd0b-4a7b-b176-5cedaa7634ad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.088295] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] ab036e1f-a62a-490a-818c-2e1256c781f6/ab036e1f-a62a-490a-818c-2e1256c781f6.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 914.089752] env[62914]: DEBUG nova.policy [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80089586cfda4cd2bdfc24d9f556239c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09d51090ac5d41fc966c7e132f288824', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 914.091789] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1588efc-6d68-4649-9bc5-feaa7667f2f3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.112410] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 914.112410] env[62914]: value = "task-1352734" [ 914.112410] env[62914]: _type = "Task" [ 914.112410] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.120266] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352734, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.315015] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Releasing lock "refresh_cache-b8aebca4-2724-419e-aaf7-d180e299a459" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.315015] env[62914]: DEBUG nova.compute.manager [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Instance network_info: |[{"id": "ad5c411a-d657-4b8e-8cb7-1ef6d90a0bc2", "address": "fa:16:3e:a1:36:fa", "network": {"id": "f59afc5a-2dbe-4ddd-962e-6ec7eb665075", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-240973820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ce1d1c0e5e04202adf074388b43266f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "305ccd93-08cb-4658-845c-d9b64952daf7", "external-id": "nsx-vlan-transportzone-490", "segmentation_id": 490, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad5c411a-d6", "ovs_interfaceid": "ad5c411a-d657-4b8e-8cb7-1ef6d90a0bc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 914.315015] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:36:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '305ccd93-08cb-4658-845c-d9b64952daf7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ad5c411a-d657-4b8e-8cb7-1ef6d90a0bc2', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.322799] env[62914]: DEBUG oslo.service.loopingcall [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.323448] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 914.323967] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-29fa2599-f908-42a5-83e7-eca6ffdc5a16 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.351020] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.351020] env[62914]: value = "task-1352735" [ 914.351020] env[62914]: _type = "Task" [ 914.351020] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.358036] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352735, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.407056] env[62914]: DEBUG nova.network.neutron [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Successfully created port: 615c8b1c-438d-43fb-999b-89172ade3f61 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 914.414323] env[62914]: DEBUG oslo_concurrency.lockutils [req-565bf72d-5974-46ed-8c3f-08c49abf0ecc req-1a576fed-04f2-4ecf-9aed-9c49765cb15e service nova] Releasing lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.414323] env[62914]: DEBUG nova.compute.manager [req-565bf72d-5974-46ed-8c3f-08c49abf0ecc req-1a576fed-04f2-4ecf-9aed-9c49765cb15e service nova] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Received event network-vif-deleted-9926578f-eff9-4aa1-b329-afaf16f5b7c0 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.528076] env[62914]: DEBUG nova.compute.manager [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 914.543458] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3eaa0436-f206-4397-8049-d25eb15df316 tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Lock "f14f4ab4-5ef1-4f2c-ae97-ac924a446df0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.101s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.600457] env[62914]: DEBUG nova.network.neutron [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 914.626479] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352734, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.741336] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b148a43e-d6cc-4b9f-8f3d-3627b7c83cfe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.748823] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce751d67-f770-4508-90d0-286fa8af9081 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.781646] env[62914]: DEBUG nova.network.neutron [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Updating instance_info_cache with network_info: [{"id": "d621ec78-18bc-443c-bc86-a1bd2846b28d", "address": "fa:16:3e:0c:01:32", "network": {"id": "f59afc5a-2dbe-4ddd-962e-6ec7eb665075", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-240973820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ce1d1c0e5e04202adf074388b43266f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "305ccd93-08cb-4658-845c-d9b64952daf7", "external-id": "nsx-vlan-transportzone-490", "segmentation_id": 490, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd621ec78-18", "ovs_interfaceid": "d621ec78-18bc-443c-bc86-a1bd2846b28d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.784055] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6520026d-cd84-4ecc-bddf-10c20f47c50a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.791960] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f40ee6-1938-4876-b2f4-cd4ba80c92b8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.806950] env[62914]: DEBUG nova.compute.provider_tree [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.859873] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352735, 'name': CreateVM_Task, 'duration_secs': 0.47591} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.860078] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 914.860860] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.860999] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.861353] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 914.861620] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e710c8d-4db0-4179-bd71-bc694888d9b3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.866356] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 914.866356] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]525ab618-9dd8-894e-55d0-2f4df65dbd62" [ 914.866356] env[62914]: _type = "Task" [ 914.866356] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.873858] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]525ab618-9dd8-894e-55d0-2f4df65dbd62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.123715] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352734, 'name': ReconfigVM_Task, 'duration_secs': 0.680921} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.124227] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Reconfigured VM instance instance-00000053 to attach disk [datastore2] ab036e1f-a62a-490a-818c-2e1256c781f6/ab036e1f-a62a-490a-818c-2e1256c781f6.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 915.124675] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e7a19f5-59a3-4294-a9b5-8127b6116c84 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.130513] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 915.130513] env[62914]: value = "task-1352736" [ 915.130513] env[62914]: _type = "Task" [ 915.130513] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.138096] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "c7cf71d7-b823-471f-949f-9ef2b539d972" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.138343] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352736, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.177669] env[62914]: DEBUG nova.compute.manager [req-b5dcf106-3a65-4ee4-a0be-e3f9ed831d4e req-bbf1bff6-dfe3-4f74-8b95-4215fd5bfce8 service nova] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Received event network-vif-plugged-d621ec78-18bc-443c-bc86-a1bd2846b28d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.177857] env[62914]: DEBUG oslo_concurrency.lockutils [req-b5dcf106-3a65-4ee4-a0be-e3f9ed831d4e req-bbf1bff6-dfe3-4f74-8b95-4215fd5bfce8 service nova] Acquiring lock "1f7b1609-08e0-4439-aea0-b4a2ec808780-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.178083] env[62914]: DEBUG oslo_concurrency.lockutils [req-b5dcf106-3a65-4ee4-a0be-e3f9ed831d4e req-bbf1bff6-dfe3-4f74-8b95-4215fd5bfce8 service nova] Lock "1f7b1609-08e0-4439-aea0-b4a2ec808780-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.178259] env[62914]: DEBUG oslo_concurrency.lockutils [req-b5dcf106-3a65-4ee4-a0be-e3f9ed831d4e req-bbf1bff6-dfe3-4f74-8b95-4215fd5bfce8 service nova] Lock "1f7b1609-08e0-4439-aea0-b4a2ec808780-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.178427] env[62914]: DEBUG nova.compute.manager [req-b5dcf106-3a65-4ee4-a0be-e3f9ed831d4e req-bbf1bff6-dfe3-4f74-8b95-4215fd5bfce8 service nova] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] No waiting events found dispatching network-vif-plugged-d621ec78-18bc-443c-bc86-a1bd2846b28d {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 915.178593] env[62914]: WARNING nova.compute.manager [req-b5dcf106-3a65-4ee4-a0be-e3f9ed831d4e req-bbf1bff6-dfe3-4f74-8b95-4215fd5bfce8 service nova] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Received unexpected event network-vif-plugged-d621ec78-18bc-443c-bc86-a1bd2846b28d for instance with vm_state building and task_state spawning. [ 915.178757] env[62914]: DEBUG nova.compute.manager [req-b5dcf106-3a65-4ee4-a0be-e3f9ed831d4e req-bbf1bff6-dfe3-4f74-8b95-4215fd5bfce8 service nova] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Received event network-changed-d621ec78-18bc-443c-bc86-a1bd2846b28d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.178974] env[62914]: DEBUG nova.compute.manager [req-b5dcf106-3a65-4ee4-a0be-e3f9ed831d4e req-bbf1bff6-dfe3-4f74-8b95-4215fd5bfce8 service nova] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Refreshing instance network info cache due to event network-changed-d621ec78-18bc-443c-bc86-a1bd2846b28d. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 915.179182] env[62914]: DEBUG oslo_concurrency.lockutils [req-b5dcf106-3a65-4ee4-a0be-e3f9ed831d4e req-bbf1bff6-dfe3-4f74-8b95-4215fd5bfce8 service nova] Acquiring lock "refresh_cache-1f7b1609-08e0-4439-aea0-b4a2ec808780" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.286753] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Releasing lock "refresh_cache-1f7b1609-08e0-4439-aea0-b4a2ec808780" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.287081] env[62914]: DEBUG nova.compute.manager [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Instance network_info: |[{"id": "d621ec78-18bc-443c-bc86-a1bd2846b28d", "address": "fa:16:3e:0c:01:32", "network": {"id": "f59afc5a-2dbe-4ddd-962e-6ec7eb665075", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-240973820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ce1d1c0e5e04202adf074388b43266f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "305ccd93-08cb-4658-845c-d9b64952daf7", "external-id": "nsx-vlan-transportzone-490", "segmentation_id": 490, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd621ec78-18", "ovs_interfaceid": "d621ec78-18bc-443c-bc86-a1bd2846b28d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 915.287423] env[62914]: DEBUG oslo_concurrency.lockutils [req-b5dcf106-3a65-4ee4-a0be-e3f9ed831d4e req-bbf1bff6-dfe3-4f74-8b95-4215fd5bfce8 service nova] Acquired lock "refresh_cache-1f7b1609-08e0-4439-aea0-b4a2ec808780" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.287614] env[62914]: DEBUG nova.network.neutron [req-b5dcf106-3a65-4ee4-a0be-e3f9ed831d4e req-bbf1bff6-dfe3-4f74-8b95-4215fd5bfce8 service nova] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Refreshing network info cache for port d621ec78-18bc-443c-bc86-a1bd2846b28d {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 915.288924] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:01:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '305ccd93-08cb-4658-845c-d9b64952daf7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd621ec78-18bc-443c-bc86-a1bd2846b28d', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.296761] env[62914]: DEBUG oslo.service.loopingcall [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.297835] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 915.298102] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-db99e98f-bdec-4c64-89d1-4b2edf391522 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.314189] env[62914]: DEBUG nova.scheduler.client.report [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 915.323023] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.323023] env[62914]: value = "task-1352737" [ 915.323023] env[62914]: _type = "Task" [ 915.323023] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.330578] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352737, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.377789] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]525ab618-9dd8-894e-55d0-2f4df65dbd62, 'name': SearchDatastore_Task, 'duration_secs': 0.041393} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.378110] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.378356] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 915.378588] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.378737] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.378948] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.379231] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-05ed0e17-fc0b-421a-97fa-41781dace6c7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.388451] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.388451] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 915.388950] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca87e88c-c888-418c-bb42-91cbce69825e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.394284] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 915.394284] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]529b25b0-f4e4-a7c6-0907-5d393b2f1df4" [ 915.394284] env[62914]: _type = "Task" [ 915.394284] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.402671] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]529b25b0-f4e4-a7c6-0907-5d393b2f1df4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.444018] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Acquiring lock "f14f4ab4-5ef1-4f2c-ae97-ac924a446df0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.444370] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Lock "f14f4ab4-5ef1-4f2c-ae97-ac924a446df0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.444652] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Acquiring lock "f14f4ab4-5ef1-4f2c-ae97-ac924a446df0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.444909] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Lock "f14f4ab4-5ef1-4f2c-ae97-ac924a446df0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.445164] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Lock "f14f4ab4-5ef1-4f2c-ae97-ac924a446df0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.447390] env[62914]: INFO nova.compute.manager [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Terminating instance [ 915.449955] env[62914]: DEBUG nova.compute.manager [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 915.450232] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 915.451166] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a75d72-6d8b-4709-a12e-2ded137eab96 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.458645] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 915.458877] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5422788a-193d-4fc1-8a61-263a9e4a1dec {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.464541] env[62914]: DEBUG oslo_vmware.api [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Waiting for the task: (returnval){ [ 915.464541] env[62914]: value = "task-1352738" [ 915.464541] env[62914]: _type = "Task" [ 915.464541] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.473804] env[62914]: DEBUG oslo_vmware.api [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352738, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.539813] env[62914]: DEBUG nova.compute.manager [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 915.567594] env[62914]: DEBUG nova.virt.hardware [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 915.567825] env[62914]: DEBUG nova.virt.hardware [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 915.567992] env[62914]: DEBUG nova.virt.hardware [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 915.568197] env[62914]: DEBUG nova.virt.hardware [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 915.568347] env[62914]: DEBUG nova.virt.hardware [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 915.568646] env[62914]: DEBUG nova.virt.hardware [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 915.568722] env[62914]: DEBUG nova.virt.hardware [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 915.568947] env[62914]: DEBUG nova.virt.hardware [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 915.569158] env[62914]: DEBUG nova.virt.hardware [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 915.569428] env[62914]: DEBUG nova.virt.hardware [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 915.569620] env[62914]: DEBUG nova.virt.hardware [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 915.570528] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a432199-d16c-4deb-986a-346fe20785e0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.578209] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abff40cb-a2a8-41ab-8ee8-2f233550068f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.642888] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352736, 'name': Rename_Task, 'duration_secs': 0.176383} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.643267] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 915.643499] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98de14fc-77df-460f-8f1c-17cdc37e03d6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.650113] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 915.650113] env[62914]: value = "task-1352739" [ 915.650113] env[62914]: _type = "Task" [ 915.650113] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.660486] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352739, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.802632] env[62914]: DEBUG nova.compute.manager [req-4984e635-fa0f-4bef-94c1-da0bf816e65e req-9cca2861-a744-40d3-ac43-39d6b8dbc3cd service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Received event network-vif-plugged-615c8b1c-438d-43fb-999b-89172ade3f61 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.802831] env[62914]: DEBUG oslo_concurrency.lockutils [req-4984e635-fa0f-4bef-94c1-da0bf816e65e req-9cca2861-a744-40d3-ac43-39d6b8dbc3cd service nova] Acquiring lock "1e527707-758c-4006-af83-9c739b9645ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.803059] env[62914]: DEBUG oslo_concurrency.lockutils [req-4984e635-fa0f-4bef-94c1-da0bf816e65e req-9cca2861-a744-40d3-ac43-39d6b8dbc3cd service nova] Lock "1e527707-758c-4006-af83-9c739b9645ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.803287] env[62914]: DEBUG oslo_concurrency.lockutils [req-4984e635-fa0f-4bef-94c1-da0bf816e65e req-9cca2861-a744-40d3-ac43-39d6b8dbc3cd service nova] Lock "1e527707-758c-4006-af83-9c739b9645ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.803468] env[62914]: DEBUG nova.compute.manager [req-4984e635-fa0f-4bef-94c1-da0bf816e65e req-9cca2861-a744-40d3-ac43-39d6b8dbc3cd service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] No waiting events found dispatching network-vif-plugged-615c8b1c-438d-43fb-999b-89172ade3f61 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 915.803640] env[62914]: WARNING nova.compute.manager [req-4984e635-fa0f-4bef-94c1-da0bf816e65e req-9cca2861-a744-40d3-ac43-39d6b8dbc3cd service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Received unexpected event network-vif-plugged-615c8b1c-438d-43fb-999b-89172ade3f61 for instance with vm_state building and task_state spawning. [ 915.819580] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.307s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.820156] env[62914]: DEBUG nova.compute.manager [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 915.823140] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.819s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.824955] env[62914]: INFO nova.compute.claims [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.843681] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352737, 'name': CreateVM_Task, 'duration_secs': 0.333332} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.843681] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 915.844209] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.844309] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.844613] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.844869] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfcb8d52-16d1-4bc6-a3f6-79e79bb0b3f9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.851119] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 915.851119] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52787f3e-0919-4fe7-e4ca-5a66e3fb1375" [ 915.851119] env[62914]: _type = "Task" [ 915.851119] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.862696] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52787f3e-0919-4fe7-e4ca-5a66e3fb1375, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.905120] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]529b25b0-f4e4-a7c6-0907-5d393b2f1df4, 'name': SearchDatastore_Task, 'duration_secs': 0.010435} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.906024] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0a93e1a-1b00-4c60-913b-e720450d01f7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.915274] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 915.915274] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52779348-9216-67f7-027e-eb5230028648" [ 915.915274] env[62914]: _type = "Task" [ 915.915274] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.920273] env[62914]: DEBUG nova.network.neutron [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Successfully updated port: 615c8b1c-438d-43fb-999b-89172ade3f61 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 915.925340] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52779348-9216-67f7-027e-eb5230028648, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.976030] env[62914]: DEBUG oslo_vmware.api [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352738, 'name': PowerOffVM_Task, 'duration_secs': 0.199498} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.976030] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 915.976030] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 915.976030] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66e124a3-5015-4d31-b3db-107c08e91fb5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.050084] env[62914]: DEBUG nova.network.neutron [req-b5dcf106-3a65-4ee4-a0be-e3f9ed831d4e req-bbf1bff6-dfe3-4f74-8b95-4215fd5bfce8 service nova] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Updated VIF entry in instance network info cache for port d621ec78-18bc-443c-bc86-a1bd2846b28d. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 916.050515] env[62914]: DEBUG nova.network.neutron [req-b5dcf106-3a65-4ee4-a0be-e3f9ed831d4e req-bbf1bff6-dfe3-4f74-8b95-4215fd5bfce8 service nova] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Updating instance_info_cache with network_info: [{"id": "d621ec78-18bc-443c-bc86-a1bd2846b28d", "address": "fa:16:3e:0c:01:32", "network": {"id": "f59afc5a-2dbe-4ddd-962e-6ec7eb665075", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-240973820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ce1d1c0e5e04202adf074388b43266f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "305ccd93-08cb-4658-845c-d9b64952daf7", "external-id": "nsx-vlan-transportzone-490", "segmentation_id": 490, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd621ec78-18", "ovs_interfaceid": "d621ec78-18bc-443c-bc86-a1bd2846b28d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.052642] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 916.052865] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 916.053062] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Deleting the datastore file [datastore2] f14f4ab4-5ef1-4f2c-ae97-ac924a446df0 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.053565] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fdec2f02-a22c-478d-81dd-3fe655bee3a6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.061474] env[62914]: DEBUG oslo_vmware.api [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Waiting for the task: (returnval){ [ 916.061474] env[62914]: value = "task-1352741" [ 916.061474] env[62914]: _type = "Task" [ 916.061474] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.071879] env[62914]: DEBUG oslo_vmware.api [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352741, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.161376] env[62914]: DEBUG oslo_vmware.api [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352739, 'name': PowerOnVM_Task, 'duration_secs': 0.44931} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.161679] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 916.161871] env[62914]: INFO nova.compute.manager [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Took 7.58 seconds to spawn the instance on the hypervisor. [ 916.162113] env[62914]: DEBUG nova.compute.manager [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.162891] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd44155-b8d5-42b6-9639-d143ca6a2cbb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.332210] env[62914]: DEBUG nova.compute.utils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 916.333651] env[62914]: DEBUG nova.compute.manager [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 916.333826] env[62914]: DEBUG nova.network.neutron [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 916.361276] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52787f3e-0919-4fe7-e4ca-5a66e3fb1375, 'name': SearchDatastore_Task, 'duration_secs': 0.011272} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.361578] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.361809] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 916.362094] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.362251] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.362435] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 916.362694] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7adf51c9-e111-407a-a3b7-cac6888148c4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.371067] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 916.371265] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 916.373053] env[62914]: DEBUG nova.policy [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce06463e715b4975bc43b14be64ef9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '619be7b3bf1445b68ba7adefc98f8782', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 916.374378] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1280e3b2-5987-4f4c-b786-16163bfdde19 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.379443] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 916.379443] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52af4cee-b0af-521d-e929-83c9828c180c" [ 916.379443] env[62914]: _type = "Task" [ 916.379443] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.387185] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52af4cee-b0af-521d-e929-83c9828c180c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.425198] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52779348-9216-67f7-027e-eb5230028648, 'name': SearchDatastore_Task, 'duration_secs': 0.011327} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.425691] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.425821] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.425966] env[62914]: DEBUG nova.network.neutron [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 916.427056] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.427337] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] b8aebca4-2724-419e-aaf7-d180e299a459/b8aebca4-2724-419e-aaf7-d180e299a459.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 916.427817] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49382d73-308a-48e6-a35a-f378b85200b3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.435561] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 916.435561] env[62914]: value = "task-1352742" [ 916.435561] env[62914]: _type = "Task" [ 916.435561] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.445227] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352742, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.555280] env[62914]: DEBUG oslo_concurrency.lockutils [req-b5dcf106-3a65-4ee4-a0be-e3f9ed831d4e req-bbf1bff6-dfe3-4f74-8b95-4215fd5bfce8 service nova] Releasing lock "refresh_cache-1f7b1609-08e0-4439-aea0-b4a2ec808780" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.572560] env[62914]: DEBUG oslo_vmware.api [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Task: {'id': task-1352741, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.430072} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.572839] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 916.573045] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 916.573234] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 916.573414] env[62914]: INFO nova.compute.manager [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Took 1.12 seconds to destroy the instance on the hypervisor. [ 916.573657] env[62914]: DEBUG oslo.service.loopingcall [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.573857] env[62914]: DEBUG nova.compute.manager [-] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 916.573954] env[62914]: DEBUG nova.network.neutron [-] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 916.652961] env[62914]: DEBUG nova.network.neutron [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Successfully created port: 160f0b55-ba33-4aa2-9d75-2c935fa56fc4 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.679699] env[62914]: INFO nova.compute.manager [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Took 17.50 seconds to build instance. [ 916.838554] env[62914]: DEBUG nova.compute.manager [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 916.893674] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52af4cee-b0af-521d-e929-83c9828c180c, 'name': SearchDatastore_Task, 'duration_secs': 0.009489} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.894554] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a977ef7-6b86-4182-bde1-35b034f8f8dd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.903582] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 916.903582] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52182a25-08e9-3474-d801-f7f3a189330e" [ 916.903582] env[62914]: _type = "Task" [ 916.903582] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.916577] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52182a25-08e9-3474-d801-f7f3a189330e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.950395] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352742, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.963225] env[62914]: DEBUG nova.network.neutron [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 917.076419] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf2a31a-0841-4b92-96f5-54326b8d4d93 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.084586] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b48201c-cdb5-4486-96d6-a3956c2c01a2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.119364] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a447e46c-9093-4a96-8ef6-d3ca53db2514 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.128481] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-975fd65f-486c-4ae2-a795-4e558434b9ab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.142177] env[62914]: DEBUG nova.compute.provider_tree [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.154953] env[62914]: DEBUG nova.network.neutron [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updating instance_info_cache with network_info: [{"id": "615c8b1c-438d-43fb-999b-89172ade3f61", "address": "fa:16:3e:9d:b8:15", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap615c8b1c-43", "ovs_interfaceid": "615c8b1c-438d-43fb-999b-89172ade3f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.182634] env[62914]: DEBUG oslo_concurrency.lockutils [None req-cdd161eb-33b7-46ed-b808-a6aeb9cc3ff7 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "ab036e1f-a62a-490a-818c-2e1256c781f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.024s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.347089] env[62914]: DEBUG nova.network.neutron [-] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.413855] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52182a25-08e9-3474-d801-f7f3a189330e, 'name': SearchDatastore_Task, 'duration_secs': 0.027308} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.414129] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.414421] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 1f7b1609-08e0-4439-aea0-b4a2ec808780/1f7b1609-08e0-4439-aea0-b4a2ec808780.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 917.414649] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9157971-74db-4d99-b3ab-9e28708751da {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.421511] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 917.421511] env[62914]: value = "task-1352743" [ 917.421511] env[62914]: _type = "Task" [ 917.421511] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.428691] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352743, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.447290] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352742, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.721247} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.447533] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] b8aebca4-2724-419e-aaf7-d180e299a459/b8aebca4-2724-419e-aaf7-d180e299a459.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 917.447746] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 917.447987] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f0410cd-cded-484e-8169-2b84833e8189 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.454278] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 917.454278] env[62914]: value = "task-1352744" [ 917.454278] env[62914]: _type = "Task" [ 917.454278] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.461506] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352744, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.645082] env[62914]: DEBUG nova.scheduler.client.report [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.658024] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.658024] env[62914]: DEBUG nova.compute.manager [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Instance network_info: |[{"id": "615c8b1c-438d-43fb-999b-89172ade3f61", "address": "fa:16:3e:9d:b8:15", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap615c8b1c-43", "ovs_interfaceid": "615c8b1c-438d-43fb-999b-89172ade3f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 917.658024] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:b8:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e31264e2-3e0a-4dfb-ba1f-6389d7d47548', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '615c8b1c-438d-43fb-999b-89172ade3f61', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.665450] env[62914]: DEBUG oslo.service.loopingcall [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.667117] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 917.667117] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f90318bc-15df-49d8-9f38-aedd916389ea {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.688420] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.688420] env[62914]: value = "task-1352745" [ 917.688420] env[62914]: _type = "Task" [ 917.688420] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.696536] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352745, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.833237] env[62914]: DEBUG nova.compute.manager [req-380691bf-a4b0-4ae9-ad73-622644c8864b req-e6c14750-e890-470c-ab95-7cf1ed1417c9 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Received event network-changed-615c8b1c-438d-43fb-999b-89172ade3f61 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.833457] env[62914]: DEBUG nova.compute.manager [req-380691bf-a4b0-4ae9-ad73-622644c8864b req-e6c14750-e890-470c-ab95-7cf1ed1417c9 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Refreshing instance network info cache due to event network-changed-615c8b1c-438d-43fb-999b-89172ade3f61. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 917.833711] env[62914]: DEBUG oslo_concurrency.lockutils [req-380691bf-a4b0-4ae9-ad73-622644c8864b req-e6c14750-e890-470c-ab95-7cf1ed1417c9 service nova] Acquiring lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.834316] env[62914]: DEBUG oslo_concurrency.lockutils [req-380691bf-a4b0-4ae9-ad73-622644c8864b req-e6c14750-e890-470c-ab95-7cf1ed1417c9 service nova] Acquired lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.834316] env[62914]: DEBUG nova.network.neutron [req-380691bf-a4b0-4ae9-ad73-622644c8864b req-e6c14750-e890-470c-ab95-7cf1ed1417c9 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Refreshing network info cache for port 615c8b1c-438d-43fb-999b-89172ade3f61 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 917.850034] env[62914]: DEBUG nova.compute.manager [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 917.851924] env[62914]: INFO nova.compute.manager [-] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Took 1.28 seconds to deallocate network for instance. [ 917.885731] env[62914]: DEBUG nova.virt.hardware [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.886034] env[62914]: DEBUG nova.virt.hardware [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.886218] env[62914]: DEBUG nova.virt.hardware [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.886411] env[62914]: DEBUG nova.virt.hardware [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.886565] env[62914]: DEBUG nova.virt.hardware [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.886721] env[62914]: DEBUG nova.virt.hardware [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.886951] env[62914]: DEBUG nova.virt.hardware [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.887164] env[62914]: DEBUG nova.virt.hardware [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.887352] env[62914]: DEBUG nova.virt.hardware [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.887526] env[62914]: DEBUG nova.virt.hardware [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.887723] env[62914]: DEBUG nova.virt.hardware [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.888661] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc4dad2b-ace6-4aae-86fc-ec687c0eda66 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.898700] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf1d40d-b233-440d-ab20-739977ad19bf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.932993] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352743, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.968325] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352744, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067168} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.968614] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.969551] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13b21f1-44c9-4a12-8620-0460866c9620 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.994330] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] b8aebca4-2724-419e-aaf7-d180e299a459/b8aebca4-2724-419e-aaf7-d180e299a459.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.994693] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df053e9c-670d-47c4-a71b-536b59ca5e6c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.017866] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 918.017866] env[62914]: value = "task-1352746" [ 918.017866] env[62914]: _type = "Task" [ 918.017866] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.027553] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352746, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.150422] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.327s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.151265] env[62914]: DEBUG nova.compute.manager [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 918.155292] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.227s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.155575] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.157924] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.493s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.158212] env[62914]: DEBUG nova.objects.instance [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lazy-loading 'resources' on Instance uuid 1fe15423-f8be-4763-b55e-2cbb383ff01d {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.188995] env[62914]: INFO nova.scheduler.client.report [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Deleted allocations for instance 4bf76ce0-b5d2-4184-888c-d0ef39878356 [ 918.212899] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352745, 'name': CreateVM_Task, 'duration_secs': 0.443759} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.214168] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 918.217273] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.217273] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.217273] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 918.217273] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54d75513-373d-443b-b9d2-16c03fc76933 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.222751] env[62914]: DEBUG nova.compute.manager [req-44f4c4f2-7c3e-4f9b-8699-d248b3ccd1e1 req-8dd63d8d-97b3-4559-8b9f-7d0d2b9c0e95 service nova] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Received event network-vif-plugged-160f0b55-ba33-4aa2-9d75-2c935fa56fc4 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.222977] env[62914]: DEBUG oslo_concurrency.lockutils [req-44f4c4f2-7c3e-4f9b-8699-d248b3ccd1e1 req-8dd63d8d-97b3-4559-8b9f-7d0d2b9c0e95 service nova] Acquiring lock "f616eac4-7c75-484e-bb64-5e227a08df8c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.223250] env[62914]: DEBUG oslo_concurrency.lockutils [req-44f4c4f2-7c3e-4f9b-8699-d248b3ccd1e1 req-8dd63d8d-97b3-4559-8b9f-7d0d2b9c0e95 service nova] Lock "f616eac4-7c75-484e-bb64-5e227a08df8c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.223564] env[62914]: DEBUG oslo_concurrency.lockutils [req-44f4c4f2-7c3e-4f9b-8699-d248b3ccd1e1 req-8dd63d8d-97b3-4559-8b9f-7d0d2b9c0e95 service nova] Lock "f616eac4-7c75-484e-bb64-5e227a08df8c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.223621] env[62914]: DEBUG nova.compute.manager [req-44f4c4f2-7c3e-4f9b-8699-d248b3ccd1e1 req-8dd63d8d-97b3-4559-8b9f-7d0d2b9c0e95 service nova] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] No waiting events found dispatching network-vif-plugged-160f0b55-ba33-4aa2-9d75-2c935fa56fc4 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.223793] env[62914]: WARNING nova.compute.manager [req-44f4c4f2-7c3e-4f9b-8699-d248b3ccd1e1 req-8dd63d8d-97b3-4559-8b9f-7d0d2b9c0e95 service nova] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Received unexpected event network-vif-plugged-160f0b55-ba33-4aa2-9d75-2c935fa56fc4 for instance with vm_state building and task_state spawning. [ 918.226459] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 918.226459] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52cb975b-c537-7639-26c7-bc48b6661368" [ 918.226459] env[62914]: _type = "Task" [ 918.226459] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.240327] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52cb975b-c537-7639-26c7-bc48b6661368, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.344953] env[62914]: DEBUG nova.network.neutron [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Successfully updated port: 160f0b55-ba33-4aa2-9d75-2c935fa56fc4 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.359198] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.432387] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352743, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.874502} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.432658] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 1f7b1609-08e0-4439-aea0-b4a2ec808780/1f7b1609-08e0-4439-aea0-b4a2ec808780.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 918.432888] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 918.433153] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-92087329-99fb-4057-8cee-03767e996200 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.442351] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 918.442351] env[62914]: value = "task-1352747" [ 918.442351] env[62914]: _type = "Task" [ 918.442351] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.450296] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352747, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.528112] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352746, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.531229] env[62914]: DEBUG nova.network.neutron [req-380691bf-a4b0-4ae9-ad73-622644c8864b req-e6c14750-e890-470c-ab95-7cf1ed1417c9 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updated VIF entry in instance network info cache for port 615c8b1c-438d-43fb-999b-89172ade3f61. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 918.531570] env[62914]: DEBUG nova.network.neutron [req-380691bf-a4b0-4ae9-ad73-622644c8864b req-e6c14750-e890-470c-ab95-7cf1ed1417c9 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updating instance_info_cache with network_info: [{"id": "615c8b1c-438d-43fb-999b-89172ade3f61", "address": "fa:16:3e:9d:b8:15", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap615c8b1c-43", "ovs_interfaceid": "615c8b1c-438d-43fb-999b-89172ade3f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.661367] env[62914]: DEBUG nova.compute.utils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.665859] env[62914]: DEBUG nova.compute.manager [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 918.666013] env[62914]: DEBUG nova.network.neutron [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 918.705586] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16e419ea-ae23-4f2a-85d7-5d2dc785ae1b tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "4bf76ce0-b5d2-4184-888c-d0ef39878356" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.710s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.714579] env[62914]: DEBUG nova.policy [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '795b94761b144a1e99ed79838d9687ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9278818cebd842669b802a4e43e91774', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 918.740717] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52cb975b-c537-7639-26c7-bc48b6661368, 'name': SearchDatastore_Task, 'duration_secs': 0.024889} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.741314] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.741588] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 918.741834] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.741990] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.742193] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 918.744641] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec43cb6c-bfa0-48cc-9b19-3a2ea5c275ad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.754683] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 918.754816] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 918.757735] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4df10887-12fc-44c4-9565-caff820a6759 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.765041] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 918.765041] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c8ab9a-86fe-a93a-814d-713a3137bad0" [ 918.765041] env[62914]: _type = "Task" [ 918.765041] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.776524] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c8ab9a-86fe-a93a-814d-713a3137bad0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.842495] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ffc8e23-6c72-4f01-b29f-5bcb2940ca59 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.848279] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "refresh_cache-f616eac4-7c75-484e-bb64-5e227a08df8c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.848455] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "refresh_cache-f616eac4-7c75-484e-bb64-5e227a08df8c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.848610] env[62914]: DEBUG nova.network.neutron [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 918.853358] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83ef11e-d3bb-411f-8570-7690f6f1f29c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.887111] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715acbc4-d5f7-46bb-a727-b8d121a19fcd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.895779] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-879e45af-be8a-459b-9bd9-a8d605951777 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.909882] env[62914]: DEBUG nova.compute.provider_tree [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.952696] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352747, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077971} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.952985] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 918.953831] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-014e46a3-7a0c-4f7b-964f-3a133a16dc44 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.976502] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 1f7b1609-08e0-4439-aea0-b4a2ec808780/1f7b1609-08e0-4439-aea0-b4a2ec808780.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.976830] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9d1c332-e525-4908-91ab-6717fd377d08 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.997733] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 918.997733] env[62914]: value = "task-1352748" [ 918.997733] env[62914]: _type = "Task" [ 918.997733] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.006464] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352748, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.007264] env[62914]: DEBUG nova.network.neutron [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Successfully created port: 79e0436a-5077-4b43-8967-969e5a2e3799 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 919.027486] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352746, 'name': ReconfigVM_Task, 'duration_secs': 0.721494} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.027774] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Reconfigured VM instance instance-00000054 to attach disk [datastore2] b8aebca4-2724-419e-aaf7-d180e299a459/b8aebca4-2724-419e-aaf7-d180e299a459.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.028502] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-987b7694-d9f2-42c3-9b4b-7a4b420d282c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.034099] env[62914]: DEBUG oslo_concurrency.lockutils [req-380691bf-a4b0-4ae9-ad73-622644c8864b req-e6c14750-e890-470c-ab95-7cf1ed1417c9 service nova] Releasing lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.035036] env[62914]: DEBUG nova.compute.manager [req-380691bf-a4b0-4ae9-ad73-622644c8864b req-e6c14750-e890-470c-ab95-7cf1ed1417c9 service nova] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Received event network-vif-deleted-0639e270-95a6-45c1-80f9-a12d293758e3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.035886] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 919.035886] env[62914]: value = "task-1352749" [ 919.035886] env[62914]: _type = "Task" [ 919.035886] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.044525] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352749, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.169445] env[62914]: DEBUG nova.compute.manager [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 919.276626] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c8ab9a-86fe-a93a-814d-713a3137bad0, 'name': SearchDatastore_Task, 'duration_secs': 0.012648} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.277731] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ffddcee-219e-4ccd-a83f-05966028d11e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.284234] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 919.284234] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52bacecd-f33a-6399-3319-bf700b9a76fe" [ 919.284234] env[62914]: _type = "Task" [ 919.284234] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.295027] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52bacecd-f33a-6399-3319-bf700b9a76fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.382234] env[62914]: DEBUG nova.network.neutron [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 919.412863] env[62914]: DEBUG nova.scheduler.client.report [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.508851] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352748, 'name': ReconfigVM_Task, 'duration_secs': 0.314275} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.509149] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 1f7b1609-08e0-4439-aea0-b4a2ec808780/1f7b1609-08e0-4439-aea0-b4a2ec808780.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.509788] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-97336590-4cee-48fd-89e9-af7e363f18e8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.519882] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 919.519882] env[62914]: value = "task-1352750" [ 919.519882] env[62914]: _type = "Task" [ 919.519882] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.526617] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352750, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.545796] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352749, 'name': Rename_Task, 'duration_secs': 0.259556} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.546217] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 919.546514] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-337c8325-71e3-42b1-b197-3f8310f4bdbd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.554182] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 919.554182] env[62914]: value = "task-1352751" [ 919.554182] env[62914]: _type = "Task" [ 919.554182] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.564524] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352751, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.568067] env[62914]: DEBUG nova.network.neutron [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Updating instance_info_cache with network_info: [{"id": "160f0b55-ba33-4aa2-9d75-2c935fa56fc4", "address": "fa:16:3e:73:d5:1d", "network": {"id": "21b6ff73-feab-478c-848a-313ea0acbfc2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1273337660-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "619be7b3bf1445b68ba7adefc98f8782", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap160f0b55-ba", "ovs_interfaceid": "160f0b55-ba33-4aa2-9d75-2c935fa56fc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.795031] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52bacecd-f33a-6399-3319-bf700b9a76fe, 'name': SearchDatastore_Task, 'duration_secs': 0.019729} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.795321] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.795575] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 1e527707-758c-4006-af83-9c739b9645ed/1e527707-758c-4006-af83-9c739b9645ed.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 919.795835] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-524f9b74-4123-4793-a4a1-c28c98b60eb3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.803394] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 919.803394] env[62914]: value = "task-1352752" [ 919.803394] env[62914]: _type = "Task" [ 919.803394] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.810802] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352752, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.919437] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.761s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.921789] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.322s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.922164] env[62914]: DEBUG nova.objects.instance [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Lazy-loading 'resources' on Instance uuid ef27919a-f205-49d9-88d8-b350fcf5cfac {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.942030] env[62914]: INFO nova.scheduler.client.report [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Deleted allocations for instance 1fe15423-f8be-4763-b55e-2cbb383ff01d [ 920.028065] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352750, 'name': Rename_Task, 'duration_secs': 0.145315} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.028304] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 920.028549] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0efe1c75-7088-46fc-a0dc-863c40c1bb5d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.035739] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 920.035739] env[62914]: value = "task-1352753" [ 920.035739] env[62914]: _type = "Task" [ 920.035739] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.043322] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352753, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.062859] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352751, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.072513] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "refresh_cache-f616eac4-7c75-484e-bb64-5e227a08df8c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.072821] env[62914]: DEBUG nova.compute.manager [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Instance network_info: |[{"id": "160f0b55-ba33-4aa2-9d75-2c935fa56fc4", "address": "fa:16:3e:73:d5:1d", "network": {"id": "21b6ff73-feab-478c-848a-313ea0acbfc2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1273337660-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "619be7b3bf1445b68ba7adefc98f8782", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap160f0b55-ba", "ovs_interfaceid": "160f0b55-ba33-4aa2-9d75-2c935fa56fc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 920.073283] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:d5:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10b81051-1eb1-406b-888c-4548c470c77e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '160f0b55-ba33-4aa2-9d75-2c935fa56fc4', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 920.080900] env[62914]: DEBUG oslo.service.loopingcall [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.081199] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 920.081431] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc718b91-fada-4727-a3bd-cfaefa4cf684 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.104405] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 920.104405] env[62914]: value = "task-1352754" [ 920.104405] env[62914]: _type = "Task" [ 920.104405] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.114423] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352754, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.178390] env[62914]: DEBUG nova.compute.manager [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 920.206360] env[62914]: DEBUG nova.virt.hardware [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 920.206626] env[62914]: DEBUG nova.virt.hardware [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 920.206785] env[62914]: DEBUG nova.virt.hardware [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 920.206971] env[62914]: DEBUG nova.virt.hardware [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 920.207145] env[62914]: DEBUG nova.virt.hardware [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 920.207299] env[62914]: DEBUG nova.virt.hardware [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 920.207514] env[62914]: DEBUG nova.virt.hardware [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 920.207675] env[62914]: DEBUG nova.virt.hardware [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 920.207844] env[62914]: DEBUG nova.virt.hardware [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 920.208016] env[62914]: DEBUG nova.virt.hardware [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 920.208200] env[62914]: DEBUG nova.virt.hardware [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.209143] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a686f472-5179-4c6a-999e-023b5e69f99b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.217574] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe30b848-4086-4b4b-844a-493de960c8ce {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.248624] env[62914]: DEBUG nova.compute.manager [req-353d2fe8-7906-4737-b6d7-0a95f1fa5046 req-9699e8b0-15f5-4f15-8b1a-5462955990ec service nova] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Received event network-changed-160f0b55-ba33-4aa2-9d75-2c935fa56fc4 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.248907] env[62914]: DEBUG nova.compute.manager [req-353d2fe8-7906-4737-b6d7-0a95f1fa5046 req-9699e8b0-15f5-4f15-8b1a-5462955990ec service nova] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Refreshing instance network info cache due to event network-changed-160f0b55-ba33-4aa2-9d75-2c935fa56fc4. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 920.249065] env[62914]: DEBUG oslo_concurrency.lockutils [req-353d2fe8-7906-4737-b6d7-0a95f1fa5046 req-9699e8b0-15f5-4f15-8b1a-5462955990ec service nova] Acquiring lock "refresh_cache-f616eac4-7c75-484e-bb64-5e227a08df8c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.249435] env[62914]: DEBUG oslo_concurrency.lockutils [req-353d2fe8-7906-4737-b6d7-0a95f1fa5046 req-9699e8b0-15f5-4f15-8b1a-5462955990ec service nova] Acquired lock "refresh_cache-f616eac4-7c75-484e-bb64-5e227a08df8c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.249435] env[62914]: DEBUG nova.network.neutron [req-353d2fe8-7906-4737-b6d7-0a95f1fa5046 req-9699e8b0-15f5-4f15-8b1a-5462955990ec service nova] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Refreshing network info cache for port 160f0b55-ba33-4aa2-9d75-2c935fa56fc4 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 920.314710] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352752, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.451555] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1451cfc9-5ff4-435f-9b58-d1c85ed9ab35 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "1fe15423-f8be-4763-b55e-2cbb383ff01d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.053s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.551539] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.568890] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352751, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.630824] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352754, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.654461] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-009bf0fc-da87-4ccf-b670-db0b51c4baf7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.683570] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b718fb2a-7020-4294-a894-f945d105bc19 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.741240] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-712b9bd8-46e5-4838-8132-7bec0ee6271e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.760783] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84977ccf-a6f3-45a5-b5cb-55f1c75b0e6c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.789055] env[62914]: DEBUG nova.compute.provider_tree [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.819233] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352752, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.910546} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.819732] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 1e527707-758c-4006-af83-9c739b9645ed/1e527707-758c-4006-af83-9c739b9645ed.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 920.820125] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 920.820511] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8b726fdc-8c34-4b16-91ba-a10f302fcf30 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.836209] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 920.836209] env[62914]: value = "task-1352755" [ 920.836209] env[62914]: _type = "Task" [ 920.836209] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.844855] env[62914]: DEBUG nova.compute.manager [req-5998566e-f6f3-4b8d-a495-9613fc579649 req-f2b80450-1bcb-47b8-a38b-fb94c2ef9637 service nova] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Received event network-vif-plugged-79e0436a-5077-4b43-8967-969e5a2e3799 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.845672] env[62914]: DEBUG oslo_concurrency.lockutils [req-5998566e-f6f3-4b8d-a495-9613fc579649 req-f2b80450-1bcb-47b8-a38b-fb94c2ef9637 service nova] Acquiring lock "4bec4c88-bd1e-4cdf-9c8a-eee56088479b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.846078] env[62914]: DEBUG oslo_concurrency.lockutils [req-5998566e-f6f3-4b8d-a495-9613fc579649 req-f2b80450-1bcb-47b8-a38b-fb94c2ef9637 service nova] Lock "4bec4c88-bd1e-4cdf-9c8a-eee56088479b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.846360] env[62914]: DEBUG oslo_concurrency.lockutils [req-5998566e-f6f3-4b8d-a495-9613fc579649 req-f2b80450-1bcb-47b8-a38b-fb94c2ef9637 service nova] Lock "4bec4c88-bd1e-4cdf-9c8a-eee56088479b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.846634] env[62914]: DEBUG nova.compute.manager [req-5998566e-f6f3-4b8d-a495-9613fc579649 req-f2b80450-1bcb-47b8-a38b-fb94c2ef9637 service nova] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] No waiting events found dispatching network-vif-plugged-79e0436a-5077-4b43-8967-969e5a2e3799 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 920.846882] env[62914]: WARNING nova.compute.manager [req-5998566e-f6f3-4b8d-a495-9613fc579649 req-f2b80450-1bcb-47b8-a38b-fb94c2ef9637 service nova] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Received unexpected event network-vif-plugged-79e0436a-5077-4b43-8967-969e5a2e3799 for instance with vm_state building and task_state spawning. [ 920.851837] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352755, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.941055] env[62914]: DEBUG nova.network.neutron [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Successfully updated port: 79e0436a-5077-4b43-8967-969e5a2e3799 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 921.047990] env[62914]: DEBUG oslo_vmware.api [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352753, 'name': PowerOnVM_Task, 'duration_secs': 0.732455} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.049796] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 921.050076] env[62914]: INFO nova.compute.manager [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Took 7.76 seconds to spawn the instance on the hypervisor. [ 921.050282] env[62914]: DEBUG nova.compute.manager [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.051216] env[62914]: DEBUG nova.network.neutron [req-353d2fe8-7906-4737-b6d7-0a95f1fa5046 req-9699e8b0-15f5-4f15-8b1a-5462955990ec service nova] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Updated VIF entry in instance network info cache for port 160f0b55-ba33-4aa2-9d75-2c935fa56fc4. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 921.051588] env[62914]: DEBUG nova.network.neutron [req-353d2fe8-7906-4737-b6d7-0a95f1fa5046 req-9699e8b0-15f5-4f15-8b1a-5462955990ec service nova] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Updating instance_info_cache with network_info: [{"id": "160f0b55-ba33-4aa2-9d75-2c935fa56fc4", "address": "fa:16:3e:73:d5:1d", "network": {"id": "21b6ff73-feab-478c-848a-313ea0acbfc2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1273337660-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "619be7b3bf1445b68ba7adefc98f8782", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap160f0b55-ba", "ovs_interfaceid": "160f0b55-ba33-4aa2-9d75-2c935fa56fc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.053332] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf6e6f9-3634-4cb5-9cb6-de54d7c3835b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.066025] env[62914]: DEBUG oslo_vmware.api [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352751, 'name': PowerOnVM_Task, 'duration_secs': 1.340289} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.066887] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 921.067099] env[62914]: INFO nova.compute.manager [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Took 10.16 seconds to spawn the instance on the hypervisor. [ 921.067296] env[62914]: DEBUG nova.compute.manager [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.073221] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23500679-17f1-42f7-8b63-3afe1b1143af {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.118144] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352754, 'name': CreateVM_Task, 'duration_secs': 0.719757} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.118326] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 921.119050] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.119236] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.119568] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 921.120048] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6990b216-6a25-424e-a3ed-f95253051b74 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.125851] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 921.125851] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]525f5c24-6962-3028-ed99-d92bd9ebccc3" [ 921.125851] env[62914]: _type = "Task" [ 921.125851] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.134735] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]525f5c24-6962-3028-ed99-d92bd9ebccc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.293065] env[62914]: DEBUG nova.scheduler.client.report [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.345922] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352755, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087099} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.346364] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 921.347015] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4e8b01-06c0-4ebb-bb0b-88d11b349f01 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.369929] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 1e527707-758c-4006-af83-9c739b9645ed/1e527707-758c-4006-af83-9c739b9645ed.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.370570] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1cc89eb-3d89-4f6c-a150-ab1bd861c00b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.391282] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 921.391282] env[62914]: value = "task-1352756" [ 921.391282] env[62914]: _type = "Task" [ 921.391282] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.399286] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352756, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.444303] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "refresh_cache-4bec4c88-bd1e-4cdf-9c8a-eee56088479b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.444600] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "refresh_cache-4bec4c88-bd1e-4cdf-9c8a-eee56088479b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.445126] env[62914]: DEBUG nova.network.neutron [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 921.561578] env[62914]: DEBUG oslo_concurrency.lockutils [req-353d2fe8-7906-4737-b6d7-0a95f1fa5046 req-9699e8b0-15f5-4f15-8b1a-5462955990ec service nova] Releasing lock "refresh_cache-f616eac4-7c75-484e-bb64-5e227a08df8c" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.589818] env[62914]: INFO nova.compute.manager [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Took 21.64 seconds to build instance. [ 921.589818] env[62914]: INFO nova.compute.manager [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Took 22.02 seconds to build instance. [ 921.637524] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]525f5c24-6962-3028-ed99-d92bd9ebccc3, 'name': SearchDatastore_Task, 'duration_secs': 0.052909} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.637977] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.638329] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 921.638590] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.638891] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.639189] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.639526] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8db0b4f0-e795-47cc-9e74-4ab3db04a2e9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.653574] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.653753] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 921.654514] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a26bd07e-017b-4c6b-a5aa-0d6a4ddfb6d8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.677014] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 921.677014] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]525de22b-db41-239e-17aa-92729bd3c36e" [ 921.677014] env[62914]: _type = "Task" [ 921.677014] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.685385] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]525de22b-db41-239e-17aa-92729bd3c36e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.798360] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.876s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.800998] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.280s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.802630] env[62914]: INFO nova.compute.claims [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.831146] env[62914]: INFO nova.scheduler.client.report [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Deleted allocations for instance ef27919a-f205-49d9-88d8-b350fcf5cfac [ 921.902316] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352756, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.990080] env[62914]: DEBUG nova.network.neutron [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 922.091234] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2a5bf35-e0eb-4022-bd77-e6af75d961b4 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "1f7b1609-08e0-4439-aea0-b4a2ec808780" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.147s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.092736] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f55649fd-cc45-4a3c-87cb-d362423073ff tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "b8aebca4-2724-419e-aaf7-d180e299a459" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.537s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.184292] env[62914]: DEBUG nova.network.neutron [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Updating instance_info_cache with network_info: [{"id": "79e0436a-5077-4b43-8967-969e5a2e3799", "address": "fa:16:3e:66:b2:ae", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79e0436a-50", "ovs_interfaceid": "79e0436a-5077-4b43-8967-969e5a2e3799", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.190679] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]525de22b-db41-239e-17aa-92729bd3c36e, 'name': SearchDatastore_Task, 'duration_secs': 0.043729} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.192112] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04ffcbdc-feaf-40f5-a122-e1c00dbb0628 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.199717] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 922.199717] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5289f9c0-e477-9d38-c779-db1b854f3c32" [ 922.199717] env[62914]: _type = "Task" [ 922.199717] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.209867] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5289f9c0-e477-9d38-c779-db1b854f3c32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.341608] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f7a9bbdf-4dfc-4dfe-91d1-7c9e73b42baf tempest-ServerMetadataTestJSON-1798586568 tempest-ServerMetadataTestJSON-1798586568-project-member] Lock "ef27919a-f205-49d9-88d8-b350fcf5cfac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.734s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.402160] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352756, 'name': ReconfigVM_Task, 'duration_secs': 0.992366} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.402498] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 1e527707-758c-4006-af83-9c739b9645ed/1e527707-758c-4006-af83-9c739b9645ed.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 922.403106] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d950bf24-8282-4c5c-b00f-be7c2e237d41 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.410056] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 922.410056] env[62914]: value = "task-1352757" [ 922.410056] env[62914]: _type = "Task" [ 922.410056] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.418503] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352757, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.695681] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "refresh_cache-4bec4c88-bd1e-4cdf-9c8a-eee56088479b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.695681] env[62914]: DEBUG nova.compute.manager [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Instance network_info: |[{"id": "79e0436a-5077-4b43-8967-969e5a2e3799", "address": "fa:16:3e:66:b2:ae", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79e0436a-50", "ovs_interfaceid": "79e0436a-5077-4b43-8967-969e5a2e3799", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 922.695681] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:b2:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '32faf59b-014c-4f1f-8331-40df95bf741f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '79e0436a-5077-4b43-8967-969e5a2e3799', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.709361] env[62914]: DEBUG oslo.service.loopingcall [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.710036] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 922.715157] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-927a1f12-cf9d-4b18-b740-c0e89862fbf4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.750962] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5289f9c0-e477-9d38-c779-db1b854f3c32, 'name': SearchDatastore_Task, 'duration_secs': 0.01345} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.755031] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.755031] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] f616eac4-7c75-484e-bb64-5e227a08df8c/f616eac4-7c75-484e-bb64-5e227a08df8c.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 922.755031] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.755031] env[62914]: value = "task-1352758" [ 922.755031] env[62914]: _type = "Task" [ 922.755031] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.755031] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0995915b-9e0e-41c0-9287-1b08470a3af0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.770409] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352758, 'name': CreateVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.772705] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 922.772705] env[62914]: value = "task-1352759" [ 922.772705] env[62914]: _type = "Task" [ 922.772705] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.788627] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352759, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.825919] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "2575bc0d-cb47-4dab-a219-f84254bda47d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.826634] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "2575bc0d-cb47-4dab-a219-f84254bda47d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.879843] env[62914]: DEBUG nova.compute.manager [req-d96d2c45-e45e-49b1-b24b-98756d18a3b8 req-bbed2627-4e7d-465b-b6c0-2e4249bcd2ec service nova] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Received event network-changed-79e0436a-5077-4b43-8967-969e5a2e3799 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.880052] env[62914]: DEBUG nova.compute.manager [req-d96d2c45-e45e-49b1-b24b-98756d18a3b8 req-bbed2627-4e7d-465b-b6c0-2e4249bcd2ec service nova] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Refreshing instance network info cache due to event network-changed-79e0436a-5077-4b43-8967-969e5a2e3799. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 922.880317] env[62914]: DEBUG oslo_concurrency.lockutils [req-d96d2c45-e45e-49b1-b24b-98756d18a3b8 req-bbed2627-4e7d-465b-b6c0-2e4249bcd2ec service nova] Acquiring lock "refresh_cache-4bec4c88-bd1e-4cdf-9c8a-eee56088479b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.880505] env[62914]: DEBUG oslo_concurrency.lockutils [req-d96d2c45-e45e-49b1-b24b-98756d18a3b8 req-bbed2627-4e7d-465b-b6c0-2e4249bcd2ec service nova] Acquired lock "refresh_cache-4bec4c88-bd1e-4cdf-9c8a-eee56088479b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.880657] env[62914]: DEBUG nova.network.neutron [req-d96d2c45-e45e-49b1-b24b-98756d18a3b8 req-bbed2627-4e7d-465b-b6c0-2e4249bcd2ec service nova] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Refreshing network info cache for port 79e0436a-5077-4b43-8967-969e5a2e3799 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 922.929378] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352757, 'name': Rename_Task, 'duration_secs': 0.176227} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.932729] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 922.933248] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9091f037-e051-4b19-a153-74019ec93cc1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.941487] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 922.941487] env[62914]: value = "task-1352760" [ 922.941487] env[62914]: _type = "Task" [ 922.941487] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.951877] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352760, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.041433] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c142a3aa-4af8-427f-a430-66b0fa1f6dfe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.050754] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d52b31-9192-46d9-93cf-37c692ecf80b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.090113] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc87f13-ee9e-4b18-95f1-9d3889b558ac {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.099152] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f901d1f-5683-4887-9b9f-9177cf4a3e3f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.114404] env[62914]: DEBUG nova.compute.provider_tree [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.268526] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352758, 'name': CreateVM_Task, 'duration_secs': 0.376765} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.268713] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 923.269619] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.269869] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.270378] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.270740] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4330980-adcb-4c31-b4a1-27064ac23157 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.284181] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352759, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.285480] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 923.285480] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]528e3e36-476b-90cc-9d33-ecdbaa2d99b4" [ 923.285480] env[62914]: _type = "Task" [ 923.285480] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.300320] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]528e3e36-476b-90cc-9d33-ecdbaa2d99b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.329102] env[62914]: DEBUG nova.compute.manager [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 923.458276] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352760, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.618380] env[62914]: DEBUG nova.scheduler.client.report [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.680331] env[62914]: DEBUG nova.network.neutron [req-d96d2c45-e45e-49b1-b24b-98756d18a3b8 req-bbed2627-4e7d-465b-b6c0-2e4249bcd2ec service nova] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Updated VIF entry in instance network info cache for port 79e0436a-5077-4b43-8967-969e5a2e3799. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 923.680502] env[62914]: DEBUG nova.network.neutron [req-d96d2c45-e45e-49b1-b24b-98756d18a3b8 req-bbed2627-4e7d-465b-b6c0-2e4249bcd2ec service nova] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Updating instance_info_cache with network_info: [{"id": "79e0436a-5077-4b43-8967-969e5a2e3799", "address": "fa:16:3e:66:b2:ae", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79e0436a-50", "ovs_interfaceid": "79e0436a-5077-4b43-8967-969e5a2e3799", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.791733] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352759, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.609147} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.796256] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] f616eac4-7c75-484e-bb64-5e227a08df8c/f616eac4-7c75-484e-bb64-5e227a08df8c.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 923.796256] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 923.796256] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a7c387bf-4ea2-4ff5-a6a5-7384e3dd84db {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.806463] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]528e3e36-476b-90cc-9d33-ecdbaa2d99b4, 'name': SearchDatastore_Task, 'duration_secs': 0.052811} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.807944] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.808212] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.808453] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.808599] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.809621] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.809621] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 923.809621] env[62914]: value = "task-1352761" [ 923.809621] env[62914]: _type = "Task" [ 923.809621] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.809621] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9ba9179d-ca7d-4b43-a05d-d61c1ad85f0c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.826013] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352761, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.826013] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.826013] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 923.826013] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50a6ee43-eacc-46a4-86da-dd78caf4a5fc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.832409] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 923.832409] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52589d52-4372-ba9f-08f7-ae25ed697327" [ 923.832409] env[62914]: _type = "Task" [ 923.832409] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.849414] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52589d52-4372-ba9f-08f7-ae25ed697327, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.858771] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.955900] env[62914]: DEBUG oslo_vmware.api [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352760, 'name': PowerOnVM_Task, 'duration_secs': 0.850222} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.956632] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 923.956632] env[62914]: INFO nova.compute.manager [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Took 8.42 seconds to spawn the instance on the hypervisor. [ 923.956822] env[62914]: DEBUG nova.compute.manager [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.958192] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f17326b-c9cf-42a5-95bb-5d5c46c214cb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.123624] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.124288] env[62914]: DEBUG nova.compute.manager [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 924.127057] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.003s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.129080] env[62914]: INFO nova.compute.claims [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 924.183699] env[62914]: DEBUG oslo_concurrency.lockutils [req-d96d2c45-e45e-49b1-b24b-98756d18a3b8 req-bbed2627-4e7d-465b-b6c0-2e4249bcd2ec service nova] Releasing lock "refresh_cache-4bec4c88-bd1e-4cdf-9c8a-eee56088479b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.322085] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352761, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.140837} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.322342] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 924.323115] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd50c62-0000-4aaf-8aa8-f105828e7507 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.345735] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] f616eac4-7c75-484e-bb64-5e227a08df8c/f616eac4-7c75-484e-bb64-5e227a08df8c.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 924.348945] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0712cf2-35f1-4f4b-bad3-694bc59142a0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.369226] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52589d52-4372-ba9f-08f7-ae25ed697327, 'name': SearchDatastore_Task, 'duration_secs': 0.023019} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.371090] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 924.371090] env[62914]: value = "task-1352762" [ 924.371090] env[62914]: _type = "Task" [ 924.371090] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.371302] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ce9ac74-170a-4e71-bc9b-2dea7deb53bf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.382027] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352762, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.383035] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 924.383035] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52cdcdd2-2e46-2e6c-881b-0cbcafaad303" [ 924.383035] env[62914]: _type = "Task" [ 924.383035] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.393572] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52cdcdd2-2e46-2e6c-881b-0cbcafaad303, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.480559] env[62914]: INFO nova.compute.manager [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Took 24.16 seconds to build instance. [ 924.498139] env[62914]: DEBUG oslo_concurrency.lockutils [None req-061c184c-ec9d-408c-ad05-222d3ccedabe tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "ab036e1f-a62a-490a-818c-2e1256c781f6" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.498421] env[62914]: DEBUG oslo_concurrency.lockutils [None req-061c184c-ec9d-408c-ad05-222d3ccedabe tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "ab036e1f-a62a-490a-818c-2e1256c781f6" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.498606] env[62914]: DEBUG nova.compute.manager [None req-061c184c-ec9d-408c-ad05-222d3ccedabe tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.499637] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c81bbeb-76f9-432f-88b8-b0f2dc11d50c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.509530] env[62914]: DEBUG nova.compute.manager [None req-061c184c-ec9d-408c-ad05-222d3ccedabe tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62914) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 924.510107] env[62914]: DEBUG nova.objects.instance [None req-061c184c-ec9d-408c-ad05-222d3ccedabe tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lazy-loading 'flavor' on Instance uuid ab036e1f-a62a-490a-818c-2e1256c781f6 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.635752] env[62914]: DEBUG nova.compute.utils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 924.637335] env[62914]: DEBUG nova.compute.manager [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 924.637491] env[62914]: DEBUG nova.network.neutron [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 924.689441] env[62914]: DEBUG nova.policy [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2c21c59e39b94deea6c3701fe971198a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1dcda44f0d6a455bb17966e7c5061190', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 924.884421] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352762, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.893315] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52cdcdd2-2e46-2e6c-881b-0cbcafaad303, 'name': SearchDatastore_Task, 'duration_secs': 0.020542} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.893959] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.893959] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 4bec4c88-bd1e-4cdf-9c8a-eee56088479b/4bec4c88-bd1e-4cdf-9c8a-eee56088479b.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 924.894195] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9af31d50-e608-4b0a-8a11-cbb54d3e2972 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.902022] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 924.902022] env[62914]: value = "task-1352763" [ 924.902022] env[62914]: _type = "Task" [ 924.902022] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.909704] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352763, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.946378] env[62914]: DEBUG nova.network.neutron [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Successfully created port: 4e0349f9-75d4-4048-a903-03f627627c22 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 924.983334] env[62914]: DEBUG oslo_concurrency.lockutils [None req-867cc33d-8324-4895-9420-2416beb4d593 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "1e527707-758c-4006-af83-9c739b9645ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.676s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.015714] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-061c184c-ec9d-408c-ad05-222d3ccedabe tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.015996] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-271392be-b290-4cdc-a8e6-a02c83863a3f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.024917] env[62914]: DEBUG oslo_vmware.api [None req-061c184c-ec9d-408c-ad05-222d3ccedabe tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 925.024917] env[62914]: value = "task-1352764" [ 925.024917] env[62914]: _type = "Task" [ 925.024917] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.034718] env[62914]: DEBUG oslo_vmware.api [None req-061c184c-ec9d-408c-ad05-222d3ccedabe tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352764, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.141583] env[62914]: DEBUG nova.compute.manager [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 925.336477] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a3574e-6742-4373-9118-118b935c4bc1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.346949] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a2a040-10a8-40b0-90ba-ead19dd47095 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.390378] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce4b74f7-20f9-4dac-b918-e5fde5fc8d7f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.399745] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352762, 'name': ReconfigVM_Task, 'duration_secs': 0.825553} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.403733] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Reconfigured VM instance instance-00000057 to attach disk [datastore1] f616eac4-7c75-484e-bb64-5e227a08df8c/f616eac4-7c75-484e-bb64-5e227a08df8c.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 925.403733] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17a1ff4c-1f09-4619-aa79-24a8b34746fc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.406563] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c025b57-8274-42f7-98c1-b527b34da1e4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.426502] env[62914]: DEBUG nova.compute.provider_tree [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.433553] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352763, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.434209] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 925.434209] env[62914]: value = "task-1352765" [ 925.434209] env[62914]: _type = "Task" [ 925.434209] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.446953] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352765, 'name': Rename_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.536559] env[62914]: DEBUG oslo_vmware.api [None req-061c184c-ec9d-408c-ad05-222d3ccedabe tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352764, 'name': PowerOffVM_Task, 'duration_secs': 0.186573} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.536890] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-061c184c-ec9d-408c-ad05-222d3ccedabe tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 925.537024] env[62914]: DEBUG nova.compute.manager [None req-061c184c-ec9d-408c-ad05-222d3ccedabe tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.537807] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62415b53-07a7-4ca8-9dce-1de14b7d1d06 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.807939] env[62914]: DEBUG nova.compute.manager [req-4bc66e7e-fe62-495d-995a-089b9403abad req-3304eac1-99ed-4cd7-9289-4b666943b239 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Received event network-changed-615c8b1c-438d-43fb-999b-89172ade3f61 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.808180] env[62914]: DEBUG nova.compute.manager [req-4bc66e7e-fe62-495d-995a-089b9403abad req-3304eac1-99ed-4cd7-9289-4b666943b239 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Refreshing instance network info cache due to event network-changed-615c8b1c-438d-43fb-999b-89172ade3f61. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 925.808413] env[62914]: DEBUG oslo_concurrency.lockutils [req-4bc66e7e-fe62-495d-995a-089b9403abad req-3304eac1-99ed-4cd7-9289-4b666943b239 service nova] Acquiring lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.808564] env[62914]: DEBUG oslo_concurrency.lockutils [req-4bc66e7e-fe62-495d-995a-089b9403abad req-3304eac1-99ed-4cd7-9289-4b666943b239 service nova] Acquired lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.809514] env[62914]: DEBUG nova.network.neutron [req-4bc66e7e-fe62-495d-995a-089b9403abad req-3304eac1-99ed-4cd7-9289-4b666943b239 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Refreshing network info cache for port 615c8b1c-438d-43fb-999b-89172ade3f61 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 925.917186] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352763, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.623041} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.917186] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 4bec4c88-bd1e-4cdf-9c8a-eee56088479b/4bec4c88-bd1e-4cdf-9c8a-eee56088479b.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 925.917186] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.917186] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8f0c539a-6f64-467f-be29-ec7d90966bad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.923248] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 925.923248] env[62914]: value = "task-1352766" [ 925.923248] env[62914]: _type = "Task" [ 925.923248] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.931954] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352766, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.936039] env[62914]: DEBUG nova.scheduler.client.report [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.949410] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352765, 'name': Rename_Task, 'duration_secs': 0.274342} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.949952] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 925.950672] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fddc2c53-8daf-4bcc-96cf-8ad3289a4344 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.959255] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 925.959255] env[62914]: value = "task-1352767" [ 925.959255] env[62914]: _type = "Task" [ 925.959255] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.976018] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352767, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.052466] env[62914]: DEBUG oslo_concurrency.lockutils [None req-061c184c-ec9d-408c-ad05-222d3ccedabe tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "ab036e1f-a62a-490a-818c-2e1256c781f6" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.554s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.151440] env[62914]: DEBUG nova.compute.manager [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 926.177921] env[62914]: DEBUG nova.virt.hardware [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 926.178207] env[62914]: DEBUG nova.virt.hardware [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 926.178368] env[62914]: DEBUG nova.virt.hardware [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.178556] env[62914]: DEBUG nova.virt.hardware [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 926.178713] env[62914]: DEBUG nova.virt.hardware [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.178916] env[62914]: DEBUG nova.virt.hardware [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 926.179128] env[62914]: DEBUG nova.virt.hardware [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 926.179268] env[62914]: DEBUG nova.virt.hardware [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 926.179456] env[62914]: DEBUG nova.virt.hardware [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 926.179619] env[62914]: DEBUG nova.virt.hardware [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 926.179791] env[62914]: DEBUG nova.virt.hardware [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.180651] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8774072-fa22-4f88-a3fa-884b03bf7d80 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.189154] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962c89a6-a450-42db-bb43-967221da842b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.434036] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352766, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.193895} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.434386] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 926.435191] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d10efb-8857-462f-ae67-7aa7124b002f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.451795] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.325s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.452352] env[62914]: DEBUG nova.compute.manager [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 926.463201] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 4bec4c88-bd1e-4cdf-9c8a-eee56088479b/4bec4c88-bd1e-4cdf-9c8a-eee56088479b.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.463811] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.098s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.464176] env[62914]: DEBUG nova.objects.instance [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lazy-loading 'resources' on Instance uuid c7cf71d7-b823-471f-949f-9ef2b539d972 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.465253] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb7c4b87-d38f-4b98-a67a-decb1096cfcf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.485271] env[62914]: DEBUG nova.objects.instance [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lazy-loading 'numa_topology' on Instance uuid c7cf71d7-b823-471f-949f-9ef2b539d972 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.497197] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352767, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.499089] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 926.499089] env[62914]: value = "task-1352768" [ 926.499089] env[62914]: _type = "Task" [ 926.499089] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.509024] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352768, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.645423] env[62914]: DEBUG nova.network.neutron [req-4bc66e7e-fe62-495d-995a-089b9403abad req-3304eac1-99ed-4cd7-9289-4b666943b239 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updated VIF entry in instance network info cache for port 615c8b1c-438d-43fb-999b-89172ade3f61. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 926.645763] env[62914]: DEBUG nova.network.neutron [req-4bc66e7e-fe62-495d-995a-089b9403abad req-3304eac1-99ed-4cd7-9289-4b666943b239 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updating instance_info_cache with network_info: [{"id": "615c8b1c-438d-43fb-999b-89172ade3f61", "address": "fa:16:3e:9d:b8:15", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap615c8b1c-43", "ovs_interfaceid": "615c8b1c-438d-43fb-999b-89172ade3f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.677309] env[62914]: DEBUG nova.compute.manager [req-9bcd6441-dd3e-436b-9ae8-df1647d4b0cf req-ac6a6c26-eba7-4285-b9fd-040774c64cd7 service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Received event network-vif-plugged-4e0349f9-75d4-4048-a903-03f627627c22 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.677562] env[62914]: DEBUG oslo_concurrency.lockutils [req-9bcd6441-dd3e-436b-9ae8-df1647d4b0cf req-ac6a6c26-eba7-4285-b9fd-040774c64cd7 service nova] Acquiring lock "13954007-b5ee-4254-b99e-5fbb2f6e0add-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.677783] env[62914]: DEBUG oslo_concurrency.lockutils [req-9bcd6441-dd3e-436b-9ae8-df1647d4b0cf req-ac6a6c26-eba7-4285-b9fd-040774c64cd7 service nova] Lock "13954007-b5ee-4254-b99e-5fbb2f6e0add-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.677954] env[62914]: DEBUG oslo_concurrency.lockutils [req-9bcd6441-dd3e-436b-9ae8-df1647d4b0cf req-ac6a6c26-eba7-4285-b9fd-040774c64cd7 service nova] Lock "13954007-b5ee-4254-b99e-5fbb2f6e0add-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.678136] env[62914]: DEBUG nova.compute.manager [req-9bcd6441-dd3e-436b-9ae8-df1647d4b0cf req-ac6a6c26-eba7-4285-b9fd-040774c64cd7 service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] No waiting events found dispatching network-vif-plugged-4e0349f9-75d4-4048-a903-03f627627c22 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 926.678346] env[62914]: WARNING nova.compute.manager [req-9bcd6441-dd3e-436b-9ae8-df1647d4b0cf req-ac6a6c26-eba7-4285-b9fd-040774c64cd7 service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Received unexpected event network-vif-plugged-4e0349f9-75d4-4048-a903-03f627627c22 for instance with vm_state building and task_state spawning. [ 926.698133] env[62914]: DEBUG nova.network.neutron [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Successfully updated port: 4e0349f9-75d4-4048-a903-03f627627c22 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 926.966359] env[62914]: DEBUG nova.compute.utils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.968140] env[62914]: DEBUG nova.compute.manager [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 926.970297] env[62914]: DEBUG nova.network.neutron [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 926.983513] env[62914]: DEBUG oslo_vmware.api [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352767, 'name': PowerOnVM_Task, 'duration_secs': 0.972483} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.983793] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 926.984085] env[62914]: INFO nova.compute.manager [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Took 9.13 seconds to spawn the instance on the hypervisor. [ 926.984702] env[62914]: DEBUG nova.compute.manager [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.985184] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f589db07-4162-4c72-bbe5-fc1cdf286298 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.992039] env[62914]: DEBUG nova.objects.base [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62914) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 927.009155] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352768, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.022569] env[62914]: DEBUG nova.policy [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41c19b40819c49d8a9dacdb74ad5899b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f5fa79f0260e461c952301f904101f79', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 927.148629] env[62914]: DEBUG oslo_concurrency.lockutils [req-4bc66e7e-fe62-495d-995a-089b9403abad req-3304eac1-99ed-4cd7-9289-4b666943b239 service nova] Releasing lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.152919] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9080f7ac-3e6c-4b79-bb5c-2495718fb920 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.161017] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0350e06-d984-40ba-b6be-e6f14fd3d003 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.192936] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df10bb51-e240-47a3-a711-e9151b6e01e6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.200830] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "refresh_cache-13954007-b5ee-4254-b99e-5fbb2f6e0add" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.200830] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "refresh_cache-13954007-b5ee-4254-b99e-5fbb2f6e0add" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.200904] env[62914]: DEBUG nova.network.neutron [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 927.203135] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428a451a-36f7-4c4e-9044-2beeb733da87 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.222244] env[62914]: DEBUG nova.compute.provider_tree [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.268385] env[62914]: DEBUG nova.network.neutron [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Successfully created port: 650bc64b-7970-4ad4-8aa6-0065c010c7d4 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 927.324019] env[62914]: DEBUG nova.objects.instance [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lazy-loading 'flavor' on Instance uuid ab036e1f-a62a-490a-818c-2e1256c781f6 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.474791] env[62914]: DEBUG nova.compute.manager [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 927.508539] env[62914]: INFO nova.compute.manager [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Took 24.96 seconds to build instance. [ 927.514741] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352768, 'name': ReconfigVM_Task, 'duration_secs': 0.729305} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.515124] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 4bec4c88-bd1e-4cdf-9c8a-eee56088479b/4bec4c88-bd1e-4cdf-9c8a-eee56088479b.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.516442] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-afb8234a-63a9-4da6-bce0-73586924df5c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.529052] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 927.529052] env[62914]: value = "task-1352769" [ 927.529052] env[62914]: _type = "Task" [ 927.529052] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.538021] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352769, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.725868] env[62914]: DEBUG nova.scheduler.client.report [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.739619] env[62914]: DEBUG nova.network.neutron [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 927.828848] env[62914]: DEBUG oslo_concurrency.lockutils [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "refresh_cache-ab036e1f-a62a-490a-818c-2e1256c781f6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.829044] env[62914]: DEBUG oslo_concurrency.lockutils [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquired lock "refresh_cache-ab036e1f-a62a-490a-818c-2e1256c781f6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.829264] env[62914]: DEBUG nova.network.neutron [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 927.829457] env[62914]: DEBUG nova.objects.instance [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lazy-loading 'info_cache' on Instance uuid ab036e1f-a62a-490a-818c-2e1256c781f6 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.920857] env[62914]: DEBUG nova.network.neutron [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Updating instance_info_cache with network_info: [{"id": "4e0349f9-75d4-4048-a903-03f627627c22", "address": "fa:16:3e:d0:25:1d", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e0349f9-75", "ovs_interfaceid": "4e0349f9-75d4-4048-a903-03f627627c22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.011076] env[62914]: DEBUG oslo_concurrency.lockutils [None req-134f13ee-6f04-49fa-9047-d19fd6c4fce5 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "f616eac4-7c75-484e-bb64-5e227a08df8c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.473s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.036301] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352769, 'name': Rename_Task, 'duration_secs': 0.214194} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.036566] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 928.036827] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36b60c56-a522-4994-b7c8-69e89c2e49b8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.044076] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 928.044076] env[62914]: value = "task-1352770" [ 928.044076] env[62914]: _type = "Task" [ 928.044076] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.052909] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352770, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.232016] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.768s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.235559] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.017s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.235817] env[62914]: DEBUG nova.objects.instance [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lazy-loading 'resources' on Instance uuid aec39e18-7796-4be3-af74-478df3a78f8f {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.333672] env[62914]: DEBUG nova.objects.base [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62914) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 928.423558] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "refresh_cache-13954007-b5ee-4254-b99e-5fbb2f6e0add" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.423861] env[62914]: DEBUG nova.compute.manager [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Instance network_info: |[{"id": "4e0349f9-75d4-4048-a903-03f627627c22", "address": "fa:16:3e:d0:25:1d", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e0349f9-75", "ovs_interfaceid": "4e0349f9-75d4-4048-a903-03f627627c22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 928.424343] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:25:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a8c8175-1197-4f12-baac-ef6aba95f585', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e0349f9-75d4-4048-a903-03f627627c22', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 928.432843] env[62914]: DEBUG oslo.service.loopingcall [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.433127] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 928.433686] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-324e84cd-0a4c-4781-80b4-9976fd6f2bde {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.456769] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 928.456769] env[62914]: value = "task-1352771" [ 928.456769] env[62914]: _type = "Task" [ 928.456769] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.470979] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352771, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.485875] env[62914]: DEBUG nova.compute.manager [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 928.516307] env[62914]: DEBUG nova.virt.hardware [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 928.516725] env[62914]: DEBUG nova.virt.hardware [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 928.517012] env[62914]: DEBUG nova.virt.hardware [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 928.517334] env[62914]: DEBUG nova.virt.hardware [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 928.517596] env[62914]: DEBUG nova.virt.hardware [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 928.517855] env[62914]: DEBUG nova.virt.hardware [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 928.518213] env[62914]: DEBUG nova.virt.hardware [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 928.518495] env[62914]: DEBUG nova.virt.hardware [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 928.518804] env[62914]: DEBUG nova.virt.hardware [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 928.519091] env[62914]: DEBUG nova.virt.hardware [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 928.520030] env[62914]: DEBUG nova.virt.hardware [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 928.520763] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c6d0df-4643-4141-82a4-38876347a835 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.533085] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c9c59c-27e7-48ed-8a50-574931acfa13 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.569449] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352770, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.717294] env[62914]: DEBUG nova.compute.manager [req-7900a2c7-ed67-481c-9ffe-c218e2907d47 req-a5b01961-1a0f-43bd-a7c8-5c662c17494d service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Received event network-changed-4e0349f9-75d4-4048-a903-03f627627c22 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.717522] env[62914]: DEBUG nova.compute.manager [req-7900a2c7-ed67-481c-9ffe-c218e2907d47 req-a5b01961-1a0f-43bd-a7c8-5c662c17494d service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Refreshing instance network info cache due to event network-changed-4e0349f9-75d4-4048-a903-03f627627c22. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 928.717758] env[62914]: DEBUG oslo_concurrency.lockutils [req-7900a2c7-ed67-481c-9ffe-c218e2907d47 req-a5b01961-1a0f-43bd-a7c8-5c662c17494d service nova] Acquiring lock "refresh_cache-13954007-b5ee-4254-b99e-5fbb2f6e0add" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.718602] env[62914]: DEBUG oslo_concurrency.lockutils [req-7900a2c7-ed67-481c-9ffe-c218e2907d47 req-a5b01961-1a0f-43bd-a7c8-5c662c17494d service nova] Acquired lock "refresh_cache-13954007-b5ee-4254-b99e-5fbb2f6e0add" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.718806] env[62914]: DEBUG nova.network.neutron [req-7900a2c7-ed67-481c-9ffe-c218e2907d47 req-a5b01961-1a0f-43bd-a7c8-5c662c17494d service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Refreshing network info cache for port 4e0349f9-75d4-4048-a903-03f627627c22 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 928.748248] env[62914]: DEBUG nova.compute.manager [req-c80d565b-dee3-4252-a816-d93168da944c req-4d9b3a12-e416-4791-8268-b4e5c79e0ea2 service nova] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Received event network-vif-plugged-650bc64b-7970-4ad4-8aa6-0065c010c7d4 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.748793] env[62914]: DEBUG oslo_concurrency.lockutils [req-c80d565b-dee3-4252-a816-d93168da944c req-4d9b3a12-e416-4791-8268-b4e5c79e0ea2 service nova] Acquiring lock "412c706a-31ce-4a33-914d-5762b02b4c63-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.748793] env[62914]: DEBUG oslo_concurrency.lockutils [req-c80d565b-dee3-4252-a816-d93168da944c req-4d9b3a12-e416-4791-8268-b4e5c79e0ea2 service nova] Lock "412c706a-31ce-4a33-914d-5762b02b4c63-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.749051] env[62914]: DEBUG oslo_concurrency.lockutils [req-c80d565b-dee3-4252-a816-d93168da944c req-4d9b3a12-e416-4791-8268-b4e5c79e0ea2 service nova] Lock "412c706a-31ce-4a33-914d-5762b02b4c63-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.749554] env[62914]: DEBUG nova.compute.manager [req-c80d565b-dee3-4252-a816-d93168da944c req-4d9b3a12-e416-4791-8268-b4e5c79e0ea2 service nova] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] No waiting events found dispatching network-vif-plugged-650bc64b-7970-4ad4-8aa6-0065c010c7d4 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 928.750196] env[62914]: WARNING nova.compute.manager [req-c80d565b-dee3-4252-a816-d93168da944c req-4d9b3a12-e416-4791-8268-b4e5c79e0ea2 service nova] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Received unexpected event network-vif-plugged-650bc64b-7970-4ad4-8aa6-0065c010c7d4 for instance with vm_state building and task_state spawning. [ 928.754590] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fbcdce18-21b8-439d-b9c6-9b358ae335a2 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 34.865s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.756203] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 13.618s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.756964] env[62914]: INFO nova.compute.manager [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Unshelving [ 928.969368] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352771, 'name': CreateVM_Task, 'duration_secs': 0.452553} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.970534] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 928.971446] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd9dca28-33a7-4db4-b60f-1126eeaf2b5c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.974695] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.974898] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.975216] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 928.975500] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ba196d6-235d-4986-962b-c14f225e8c73 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.981745] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 928.981745] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]521983f7-329f-9d4e-6f2f-cfbfa632d50f" [ 928.981745] env[62914]: _type = "Task" [ 928.981745] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.990252] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb1dee8-5305-417f-874c-5e8469f7dbbe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.000355] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]521983f7-329f-9d4e-6f2f-cfbfa632d50f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.029602] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b43fc44-3751-489d-a4bd-80a44d85875d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.039073] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf547a4-0f33-47ba-8a6a-65541a3345e0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.054106] env[62914]: DEBUG nova.compute.provider_tree [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.066221] env[62914]: DEBUG oslo_vmware.api [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352770, 'name': PowerOnVM_Task, 'duration_secs': 0.850331} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.067224] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.067535] env[62914]: INFO nova.compute.manager [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Took 8.89 seconds to spawn the instance on the hypervisor. [ 929.067744] env[62914]: DEBUG nova.compute.manager [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.068565] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512e30fa-e1bc-41ff-9782-8f1ce36a449c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.109914] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "f616eac4-7c75-484e-bb64-5e227a08df8c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.110220] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "f616eac4-7c75-484e-bb64-5e227a08df8c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.110439] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "f616eac4-7c75-484e-bb64-5e227a08df8c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.110625] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "f616eac4-7c75-484e-bb64-5e227a08df8c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.110799] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "f616eac4-7c75-484e-bb64-5e227a08df8c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.113150] env[62914]: INFO nova.compute.manager [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Terminating instance [ 929.115101] env[62914]: DEBUG nova.compute.manager [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 929.115303] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 929.116146] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a1feba2-1460-41e0-8a2b-e9ad344b476e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.125923] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 929.126888] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2915085e-0581-4921-87e9-d50a2751099c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.136154] env[62914]: DEBUG oslo_vmware.api [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 929.136154] env[62914]: value = "task-1352772" [ 929.136154] env[62914]: _type = "Task" [ 929.136154] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.145967] env[62914]: DEBUG oslo_vmware.api [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352772, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.147146] env[62914]: DEBUG nova.network.neutron [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Updating instance_info_cache with network_info: [{"id": "2282e6e9-7ccf-474e-9fd5-6738ed525721", "address": "fa:16:3e:ec:98:9f", "network": {"id": "f59afc5a-2dbe-4ddd-962e-6ec7eb665075", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-240973820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ce1d1c0e5e04202adf074388b43266f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "305ccd93-08cb-4658-845c-d9b64952daf7", "external-id": "nsx-vlan-transportzone-490", "segmentation_id": 490, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2282e6e9-7c", "ovs_interfaceid": "2282e6e9-7ccf-474e-9fd5-6738ed525721", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.341578] env[62914]: DEBUG nova.network.neutron [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Successfully updated port: 650bc64b-7970-4ad4-8aa6-0065c010c7d4 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 929.439530] env[62914]: DEBUG nova.network.neutron [req-7900a2c7-ed67-481c-9ffe-c218e2907d47 req-a5b01961-1a0f-43bd-a7c8-5c662c17494d service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Updated VIF entry in instance network info cache for port 4e0349f9-75d4-4048-a903-03f627627c22. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 929.439901] env[62914]: DEBUG nova.network.neutron [req-7900a2c7-ed67-481c-9ffe-c218e2907d47 req-a5b01961-1a0f-43bd-a7c8-5c662c17494d service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Updating instance_info_cache with network_info: [{"id": "4e0349f9-75d4-4048-a903-03f627627c22", "address": "fa:16:3e:d0:25:1d", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e0349f9-75", "ovs_interfaceid": "4e0349f9-75d4-4048-a903-03f627627c22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.499519] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]521983f7-329f-9d4e-6f2f-cfbfa632d50f, 'name': SearchDatastore_Task, 'duration_secs': 0.026163} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.499762] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.500144] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.500522] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.500778] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.501094] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.501524] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e4334fd-6f82-4489-a246-5afe7226ce39 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.514887] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.515162] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 929.516612] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16d01130-230e-4059-9d16-a175d0c88104 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.523984] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 929.523984] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52248cfb-ed3d-d3db-994b-8ff3c4852bae" [ 929.523984] env[62914]: _type = "Task" [ 929.523984] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.533584] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52248cfb-ed3d-d3db-994b-8ff3c4852bae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.558751] env[62914]: DEBUG nova.scheduler.client.report [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.590063] env[62914]: INFO nova.compute.manager [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Took 26.62 seconds to build instance. [ 929.646575] env[62914]: DEBUG oslo_vmware.api [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352772, 'name': PowerOffVM_Task, 'duration_secs': 0.471366} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.646875] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 929.647057] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 929.647347] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3a213b22-a331-40f5-846f-ce2aee4c91a0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.649182] env[62914]: DEBUG oslo_concurrency.lockutils [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Releasing lock "refresh_cache-ab036e1f-a62a-490a-818c-2e1256c781f6" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.751497] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 929.751881] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 929.752121] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Deleting the datastore file [datastore1] f616eac4-7c75-484e-bb64-5e227a08df8c {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.752425] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2d067b4c-f535-47b7-87ae-84fa455054db {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.762778] env[62914]: DEBUG oslo_vmware.api [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for the task: (returnval){ [ 929.762778] env[62914]: value = "task-1352774" [ 929.762778] env[62914]: _type = "Task" [ 929.762778] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.775317] env[62914]: DEBUG oslo_vmware.api [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352774, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.779496] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.847373] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "refresh_cache-412c706a-31ce-4a33-914d-5762b02b4c63" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.847518] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "refresh_cache-412c706a-31ce-4a33-914d-5762b02b4c63" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.847701] env[62914]: DEBUG nova.network.neutron [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 929.942718] env[62914]: DEBUG oslo_concurrency.lockutils [req-7900a2c7-ed67-481c-9ffe-c218e2907d47 req-a5b01961-1a0f-43bd-a7c8-5c662c17494d service nova] Releasing lock "refresh_cache-13954007-b5ee-4254-b99e-5fbb2f6e0add" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.035442] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52248cfb-ed3d-d3db-994b-8ff3c4852bae, 'name': SearchDatastore_Task, 'duration_secs': 0.018535} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.036319] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fdefbb0-e35f-4cae-8948-33222240e237 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.042472] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 930.042472] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a1e927-4289-2a30-96b9-c980a8250e1a" [ 930.042472] env[62914]: _type = "Task" [ 930.042472] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.058668] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a1e927-4289-2a30-96b9-c980a8250e1a, 'name': SearchDatastore_Task, 'duration_secs': 0.010066} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.058959] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.059283] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 13954007-b5ee-4254-b99e-5fbb2f6e0add/13954007-b5ee-4254-b99e-5fbb2f6e0add.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 930.059577] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-503b6648-5910-43ac-9ade-ec759374a896 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.063526] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.828s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.065650] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.707s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.065889] env[62914]: DEBUG nova.objects.instance [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Lazy-loading 'resources' on Instance uuid f14f4ab4-5ef1-4f2c-ae97-ac924a446df0 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.068207] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 930.068207] env[62914]: value = "task-1352775" [ 930.068207] env[62914]: _type = "Task" [ 930.068207] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.078844] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352775, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.084765] env[62914]: INFO nova.scheduler.client.report [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleted allocations for instance aec39e18-7796-4be3-af74-478df3a78f8f [ 930.092233] env[62914]: DEBUG oslo_concurrency.lockutils [None req-82d392a4-92ac-434e-89d3-0a7d62268f97 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "4bec4c88-bd1e-4cdf-9c8a-eee56088479b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.129s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.152950] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 930.153385] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-541efd45-8cb3-4909-9c84-446782fb7ae3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.161475] env[62914]: DEBUG oslo_vmware.api [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 930.161475] env[62914]: value = "task-1352776" [ 930.161475] env[62914]: _type = "Task" [ 930.161475] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.171416] env[62914]: DEBUG oslo_vmware.api [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352776, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.274926] env[62914]: DEBUG oslo_vmware.api [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Task: {'id': task-1352774, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321441} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.275387] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 930.275668] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 930.275931] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 930.276203] env[62914]: INFO nova.compute.manager [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Took 1.16 seconds to destroy the instance on the hypervisor. [ 930.276530] env[62914]: DEBUG oslo.service.loopingcall [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.276801] env[62914]: DEBUG nova.compute.manager [-] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 930.276950] env[62914]: DEBUG nova.network.neutron [-] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 930.382593] env[62914]: DEBUG nova.network.neutron [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 930.562555] env[62914]: DEBUG nova.compute.manager [req-41e58495-3b69-4e0f-97dc-ae25067f5cef req-3b07d06c-3eca-485e-a59f-24ce69959022 service nova] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Received event network-vif-deleted-160f0b55-ba33-4aa2-9d75-2c935fa56fc4 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.562555] env[62914]: INFO nova.compute.manager [req-41e58495-3b69-4e0f-97dc-ae25067f5cef req-3b07d06c-3eca-485e-a59f-24ce69959022 service nova] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Neutron deleted interface 160f0b55-ba33-4aa2-9d75-2c935fa56fc4; detaching it from the instance and deleting it from the info cache [ 930.562711] env[62914]: DEBUG nova.network.neutron [req-41e58495-3b69-4e0f-97dc-ae25067f5cef req-3b07d06c-3eca-485e-a59f-24ce69959022 service nova] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.582748] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352775, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.593124] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a8b01fd2-4446-40e1-a7e6-b84e259a387b tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "aec39e18-7796-4be3-af74-478df3a78f8f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.035s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.597845] env[62914]: DEBUG nova.network.neutron [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Updating instance_info_cache with network_info: [{"id": "650bc64b-7970-4ad4-8aa6-0065c010c7d4", "address": "fa:16:3e:7e:3c:60", "network": {"id": "c9c30d64-cafe-48a5-b868-b20d90d8920d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-458611021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5fa79f0260e461c952301f904101f79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap650bc64b-79", "ovs_interfaceid": "650bc64b-7970-4ad4-8aa6-0065c010c7d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.605074] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "10c348bd-d85d-49b8-9b0f-216142f664d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.605369] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "10c348bd-d85d-49b8-9b0f-216142f664d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.674673] env[62914]: DEBUG oslo_vmware.api [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352776, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.742743] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a988eee3-eb6f-4334-8ddb-908081ec4069 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.748125] env[62914]: DEBUG nova.compute.manager [req-9da84e66-1d09-41e6-8737-7622d8e99131 req-c522ba71-e8b1-4a4e-a310-2cb5ebedcf23 service nova] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Received event network-changed-650bc64b-7970-4ad4-8aa6-0065c010c7d4 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.748369] env[62914]: DEBUG nova.compute.manager [req-9da84e66-1d09-41e6-8737-7622d8e99131 req-c522ba71-e8b1-4a4e-a310-2cb5ebedcf23 service nova] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Refreshing instance network info cache due to event network-changed-650bc64b-7970-4ad4-8aa6-0065c010c7d4. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 930.748521] env[62914]: DEBUG oslo_concurrency.lockutils [req-9da84e66-1d09-41e6-8737-7622d8e99131 req-c522ba71-e8b1-4a4e-a310-2cb5ebedcf23 service nova] Acquiring lock "refresh_cache-412c706a-31ce-4a33-914d-5762b02b4c63" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.754360] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397c74ab-e953-4886-a41c-4f04948ec95a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.786258] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5e0f0a-8d3e-448f-91c8-d8e55bd03813 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.796594] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6909219-c00c-4151-a3ed-360f6704aa15 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.818906] env[62914]: DEBUG nova.compute.provider_tree [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.029844] env[62914]: DEBUG nova.network.neutron [-] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.065971] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5fecf7d4-de26-44d3-afcc-54b512d2c5bc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.079716] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-178d1cba-ef91-4205-b4c3-da9033d22fab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.094368] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352775, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518166} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.095046] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 13954007-b5ee-4254-b99e-5fbb2f6e0add/13954007-b5ee-4254-b99e-5fbb2f6e0add.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 931.095286] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 931.095541] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e4a9c62c-3dda-4b06-ac5e-a1380b40d2ea {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.101327] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "refresh_cache-412c706a-31ce-4a33-914d-5762b02b4c63" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.101699] env[62914]: DEBUG nova.compute.manager [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Instance network_info: |[{"id": "650bc64b-7970-4ad4-8aa6-0065c010c7d4", "address": "fa:16:3e:7e:3c:60", "network": {"id": "c9c30d64-cafe-48a5-b868-b20d90d8920d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-458611021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5fa79f0260e461c952301f904101f79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap650bc64b-79", "ovs_interfaceid": "650bc64b-7970-4ad4-8aa6-0065c010c7d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 931.101928] env[62914]: DEBUG oslo_concurrency.lockutils [req-9da84e66-1d09-41e6-8737-7622d8e99131 req-c522ba71-e8b1-4a4e-a310-2cb5ebedcf23 service nova] Acquired lock "refresh_cache-412c706a-31ce-4a33-914d-5762b02b4c63" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.102093] env[62914]: DEBUG nova.network.neutron [req-9da84e66-1d09-41e6-8737-7622d8e99131 req-c522ba71-e8b1-4a4e-a310-2cb5ebedcf23 service nova] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Refreshing network info cache for port 650bc64b-7970-4ad4-8aa6-0065c010c7d4 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 931.103364] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:3c:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '650bc64b-7970-4ad4-8aa6-0065c010c7d4', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 931.111598] env[62914]: DEBUG oslo.service.loopingcall [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.121272] env[62914]: DEBUG nova.compute.manager [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 931.124061] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 931.125039] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 931.125039] env[62914]: value = "task-1352777" [ 931.125039] env[62914]: _type = "Task" [ 931.125039] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.125327] env[62914]: DEBUG nova.compute.manager [req-41e58495-3b69-4e0f-97dc-ae25067f5cef req-3b07d06c-3eca-485e-a59f-24ce69959022 service nova] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Detach interface failed, port_id=160f0b55-ba33-4aa2-9d75-2c935fa56fc4, reason: Instance f616eac4-7c75-484e-bb64-5e227a08df8c could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 931.126401] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd6bb535-876b-456e-9a62-1ac8758cff2f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.153329] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352777, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.154352] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 931.154352] env[62914]: value = "task-1352778" [ 931.154352] env[62914]: _type = "Task" [ 931.154352] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.163508] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352778, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.174008] env[62914]: DEBUG oslo_vmware.api [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352776, 'name': PowerOnVM_Task, 'duration_secs': 0.67297} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.174332] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 931.174545] env[62914]: DEBUG nova.compute.manager [None req-28027d6f-695b-4a2e-9101-7410b75a2064 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.175608] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232bb89a-e5af-49ed-84a1-0a0dfa8983db {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.323338] env[62914]: DEBUG nova.scheduler.client.report [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 931.532122] env[62914]: INFO nova.compute.manager [-] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Took 1.26 seconds to deallocate network for instance. [ 931.644807] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.654609] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352777, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076493} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.654991] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 931.658738] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c25131-cadc-4a2c-a8df-19e33db8845e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.681960] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 13954007-b5ee-4254-b99e-5fbb2f6e0add/13954007-b5ee-4254-b99e-5fbb2f6e0add.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.685044] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07efa3d4-e5d0-4c3b-b276-859e7c1992af {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.699752] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352778, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.712269] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 931.712269] env[62914]: value = "task-1352779" [ 931.712269] env[62914]: _type = "Task" [ 931.712269] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.723132] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352779, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.828356] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.763s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.830571] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.972s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.832193] env[62914]: INFO nova.compute.claims [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.858535] env[62914]: INFO nova.scheduler.client.report [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Deleted allocations for instance f14f4ab4-5ef1-4f2c-ae97-ac924a446df0 [ 931.916330] env[62914]: DEBUG nova.network.neutron [req-9da84e66-1d09-41e6-8737-7622d8e99131 req-c522ba71-e8b1-4a4e-a310-2cb5ebedcf23 service nova] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Updated VIF entry in instance network info cache for port 650bc64b-7970-4ad4-8aa6-0065c010c7d4. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 931.916539] env[62914]: DEBUG nova.network.neutron [req-9da84e66-1d09-41e6-8737-7622d8e99131 req-c522ba71-e8b1-4a4e-a310-2cb5ebedcf23 service nova] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Updating instance_info_cache with network_info: [{"id": "650bc64b-7970-4ad4-8aa6-0065c010c7d4", "address": "fa:16:3e:7e:3c:60", "network": {"id": "c9c30d64-cafe-48a5-b868-b20d90d8920d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-458611021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5fa79f0260e461c952301f904101f79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap650bc64b-79", "ovs_interfaceid": "650bc64b-7970-4ad4-8aa6-0065c010c7d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.040043] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.165582] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352778, 'name': CreateVM_Task, 'duration_secs': 0.779443} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.165735] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 932.166432] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.166609] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.166933] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 932.167222] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03975cd4-f42e-448e-b470-2c1a9fddac6c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.173149] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 932.173149] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]527a6124-60b0-1a2f-9509-1d06e35bdbf8" [ 932.173149] env[62914]: _type = "Task" [ 932.173149] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.182677] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]527a6124-60b0-1a2f-9509-1d06e35bdbf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.223092] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352779, 'name': ReconfigVM_Task, 'duration_secs': 0.314896} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.223407] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 13954007-b5ee-4254-b99e-5fbb2f6e0add/13954007-b5ee-4254-b99e-5fbb2f6e0add.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.224085] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be57e8ed-6cb8-474e-afd5-db32bc5a35ff {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.231343] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 932.231343] env[62914]: value = "task-1352780" [ 932.231343] env[62914]: _type = "Task" [ 932.231343] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.242671] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352780, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.366684] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3179af09-14c2-49d9-881d-a1784fe2f10f tempest-ServerAddressesTestJSON-224827980 tempest-ServerAddressesTestJSON-224827980-project-member] Lock "f14f4ab4-5ef1-4f2c-ae97-ac924a446df0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.922s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.422681] env[62914]: DEBUG oslo_concurrency.lockutils [req-9da84e66-1d09-41e6-8737-7622d8e99131 req-c522ba71-e8b1-4a4e-a310-2cb5ebedcf23 service nova] Releasing lock "refresh_cache-412c706a-31ce-4a33-914d-5762b02b4c63" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.533912] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 932.534324] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 932.534545] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Starting heal instance info cache {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 932.683938] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]527a6124-60b0-1a2f-9509-1d06e35bdbf8, 'name': SearchDatastore_Task, 'duration_secs': 0.013199} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.684274] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.684515] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 932.684761] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.684917] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.685114] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 932.685389] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d549f1cf-ec75-46c0-a361-85967732f733 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.694991] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 932.695208] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 932.695933] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83e204a0-73a4-4424-91c0-12078b9b6034 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.703050] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 932.703050] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f040af-846d-1639-fd05-10ae39a1ca7f" [ 932.703050] env[62914]: _type = "Task" [ 932.703050] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.711746] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f040af-846d-1639-fd05-10ae39a1ca7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.742122] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352780, 'name': Rename_Task, 'duration_secs': 0.161661} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.742334] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 932.742596] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f0173b2-ae2f-49a1-b500-b6fb145e0c86 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.751496] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 932.751496] env[62914]: value = "task-1352781" [ 932.751496] env[62914]: _type = "Task" [ 932.751496] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.760168] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352781, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.004578] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d7abf9-159b-41cd-9cf6-aa1356e33501 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.015575] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11f0862-7d37-4018-a8f9-93d1a3924728 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.048993] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a90abce-bec6-4de8-9460-8d464876944f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.057284] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb5dd77-1b83-40ec-b9c7-5f0429de1b44 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.073240] env[62914]: DEBUG nova.compute.provider_tree [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.215209] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f040af-846d-1639-fd05-10ae39a1ca7f, 'name': SearchDatastore_Task, 'duration_secs': 0.022962} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.216041] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bd06194-70d5-4734-96d8-1063b747f327 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.222915] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 933.222915] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c4a786-5ed7-1748-c38e-3ca8a45e69f5" [ 933.222915] env[62914]: _type = "Task" [ 933.222915] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.233210] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c4a786-5ed7-1748-c38e-3ca8a45e69f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.260683] env[62914]: DEBUG oslo_vmware.api [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352781, 'name': PowerOnVM_Task, 'duration_secs': 0.484481} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.260940] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 933.261294] env[62914]: INFO nova.compute.manager [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Took 7.11 seconds to spawn the instance on the hypervisor. [ 933.261433] env[62914]: DEBUG nova.compute.manager [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 933.262273] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a421be8d-5dd1-49d0-a1b5-a47075ce7992 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.331111] env[62914]: DEBUG oslo_concurrency.lockutils [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "1f7b1609-08e0-4439-aea0-b4a2ec808780" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.331111] env[62914]: DEBUG oslo_concurrency.lockutils [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "1f7b1609-08e0-4439-aea0-b4a2ec808780" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.331111] env[62914]: DEBUG oslo_concurrency.lockutils [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "1f7b1609-08e0-4439-aea0-b4a2ec808780-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.331111] env[62914]: DEBUG oslo_concurrency.lockutils [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "1f7b1609-08e0-4439-aea0-b4a2ec808780-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.331111] env[62914]: DEBUG oslo_concurrency.lockutils [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "1f7b1609-08e0-4439-aea0-b4a2ec808780-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.333105] env[62914]: INFO nova.compute.manager [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Terminating instance [ 933.335035] env[62914]: DEBUG nova.compute.manager [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 933.335240] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.336092] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28706c6b-55c2-495b-b187-a587991406ba {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.345140] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.348696] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8e4ce33-a00b-4459-a5e4-0fb911672f65 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.353426] env[62914]: DEBUG oslo_vmware.api [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 933.353426] env[62914]: value = "task-1352782" [ 933.353426] env[62914]: _type = "Task" [ 933.353426] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.362394] env[62914]: DEBUG oslo_vmware.api [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352782, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.577272] env[62914]: DEBUG nova.scheduler.client.report [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.734399] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c4a786-5ed7-1748-c38e-3ca8a45e69f5, 'name': SearchDatastore_Task, 'duration_secs': 0.033366} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.735078] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.735336] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 412c706a-31ce-4a33-914d-5762b02b4c63/412c706a-31ce-4a33-914d-5762b02b4c63.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 933.735707] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ff0010f3-403b-45ab-a5f2-71e6bee36ad2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.744481] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 933.744481] env[62914]: value = "task-1352783" [ 933.744481] env[62914]: _type = "Task" [ 933.744481] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.753753] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352783, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.780061] env[62914]: INFO nova.compute.manager [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Took 23.27 seconds to build instance. [ 933.863589] env[62914]: DEBUG oslo_vmware.api [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352782, 'name': PowerOffVM_Task, 'duration_secs': 0.427617} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.863855] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.864129] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 933.864586] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0e700f41-cec5-41f6-8c61-5e3fe424503c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.933543] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.933942] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.934203] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Deleting the datastore file [datastore1] 1f7b1609-08e0-4439-aea0-b4a2ec808780 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.934582] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5ce76ca0-9feb-4124-8f53-6a8f0a5b032e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.942482] env[62914]: DEBUG oslo_vmware.api [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 933.942482] env[62914]: value = "task-1352785" [ 933.942482] env[62914]: _type = "Task" [ 933.942482] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.951218] env[62914]: DEBUG oslo_vmware.api [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352785, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.084953] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.254s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.085548] env[62914]: DEBUG nova.compute.manager [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 934.088830] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.309s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.092099] env[62914]: DEBUG nova.objects.instance [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lazy-loading 'pci_requests' on Instance uuid c7cf71d7-b823-471f-949f-9ef2b539d972 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.257288] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352783, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.283162] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9b715b45-fa07-4915-bd67-8cfcc8c965e9 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "13954007-b5ee-4254-b99e-5fbb2f6e0add" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.783s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.336558] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "13954007-b5ee-4254-b99e-5fbb2f6e0add" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.336852] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "13954007-b5ee-4254-b99e-5fbb2f6e0add" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.337063] env[62914]: INFO nova.compute.manager [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Shelving [ 934.453873] env[62914]: DEBUG oslo_vmware.api [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352785, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.461499} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.454155] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.454348] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 934.454532] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 934.455468] env[62914]: INFO nova.compute.manager [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Took 1.12 seconds to destroy the instance on the hypervisor. [ 934.455468] env[62914]: DEBUG oslo.service.loopingcall [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.455468] env[62914]: DEBUG nova.compute.manager [-] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 934.455468] env[62914]: DEBUG nova.network.neutron [-] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 934.559754] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Didn't find any instances for network info cache update. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 934.560013] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 934.560270] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 934.560463] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 934.560651] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 934.560850] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 934.561008] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 934.561166] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62914) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 934.561433] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager.update_available_resource {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 934.595177] env[62914]: DEBUG nova.compute.utils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.598937] env[62914]: DEBUG nova.objects.instance [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lazy-loading 'numa_topology' on Instance uuid c7cf71d7-b823-471f-949f-9ef2b539d972 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.600049] env[62914]: DEBUG nova.compute.manager [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 934.600222] env[62914]: DEBUG nova.network.neutron [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 934.649987] env[62914]: DEBUG nova.policy [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd900131bc974504a489fcd4ffaec5b6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c3055903b6f4443bbeb9897e75e34c49', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 934.736493] env[62914]: DEBUG nova.compute.manager [req-fe5747ee-de76-4de2-98be-8e2a25a004a8 req-3fa8aac1-f450-4734-8929-41d0d3b4d8cf service nova] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Received event network-vif-deleted-d621ec78-18bc-443c-bc86-a1bd2846b28d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.736588] env[62914]: INFO nova.compute.manager [req-fe5747ee-de76-4de2-98be-8e2a25a004a8 req-3fa8aac1-f450-4734-8929-41d0d3b4d8cf service nova] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Neutron deleted interface d621ec78-18bc-443c-bc86-a1bd2846b28d; detaching it from the instance and deleting it from the info cache [ 934.736713] env[62914]: DEBUG nova.network.neutron [req-fe5747ee-de76-4de2-98be-8e2a25a004a8 req-3fa8aac1-f450-4734-8929-41d0d3b4d8cf service nova] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.757338] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352783, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681147} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.757609] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 412c706a-31ce-4a33-914d-5762b02b4c63/412c706a-31ce-4a33-914d-5762b02b4c63.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 934.757823] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 934.758098] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b00a1907-d22c-4e72-af09-d3d76d5804c8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.766446] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 934.766446] env[62914]: value = "task-1352786" [ 934.766446] env[62914]: _type = "Task" [ 934.766446] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.775563] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352786, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.846599] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.846884] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5d11091f-c842-4301-bcd6-079668ac64b0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.853548] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 934.853548] env[62914]: value = "task-1352787" [ 934.853548] env[62914]: _type = "Task" [ 934.853548] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.862201] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352787, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.937265] env[62914]: DEBUG nova.network.neutron [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Successfully created port: d41f0fe0-d70c-4f8c-a306-e14dc0107565 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 935.064288] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.104262] env[62914]: INFO nova.compute.claims [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.111128] env[62914]: DEBUG nova.compute.manager [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 935.210246] env[62914]: DEBUG nova.network.neutron [-] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.239413] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-215539ff-e16a-4b62-ab9e-741598c9ccdb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.250541] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cef9480-ca6e-4f26-a622-8c9f59e407b3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.278406] env[62914]: DEBUG nova.compute.manager [req-fe5747ee-de76-4de2-98be-8e2a25a004a8 req-3fa8aac1-f450-4734-8929-41d0d3b4d8cf service nova] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Detach interface failed, port_id=d621ec78-18bc-443c-bc86-a1bd2846b28d, reason: Instance 1f7b1609-08e0-4439-aea0-b4a2ec808780 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 935.283052] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352786, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071289} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.283052] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 935.283760] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed52bda0-95af-4c7e-b2d1-ed898fbd4194 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.305781] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 412c706a-31ce-4a33-914d-5762b02b4c63/412c706a-31ce-4a33-914d-5762b02b4c63.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 935.306041] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6512fce6-14b5-4c4f-9b63-2315652cd96d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.325784] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 935.325784] env[62914]: value = "task-1352788" [ 935.325784] env[62914]: _type = "Task" [ 935.325784] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.363486] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352787, 'name': PowerOffVM_Task, 'duration_secs': 0.230891} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.363750] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.364500] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4bd405a-9641-42f1-8296-a82dfec59252 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.383546] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936ff3d5-c735-4341-b42b-21eb98f56523 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.712037] env[62914]: INFO nova.compute.manager [-] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Took 1.26 seconds to deallocate network for instance. [ 935.837780] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352788, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.894739] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Creating Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 935.895078] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f8992ddf-9d09-4837-847d-8de6784401d3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.903214] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 935.903214] env[62914]: value = "task-1352789" [ 935.903214] env[62914]: _type = "Task" [ 935.903214] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.912244] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352789, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.120241] env[62914]: DEBUG nova.compute.manager [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 936.147085] env[62914]: DEBUG nova.virt.hardware [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.147372] env[62914]: DEBUG nova.virt.hardware [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.147537] env[62914]: DEBUG nova.virt.hardware [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.147721] env[62914]: DEBUG nova.virt.hardware [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.147872] env[62914]: DEBUG nova.virt.hardware [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.148038] env[62914]: DEBUG nova.virt.hardware [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.148264] env[62914]: DEBUG nova.virt.hardware [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.148429] env[62914]: DEBUG nova.virt.hardware [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.148597] env[62914]: DEBUG nova.virt.hardware [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.148763] env[62914]: DEBUG nova.virt.hardware [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.148936] env[62914]: DEBUG nova.virt.hardware [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.149928] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0b7e26-a33d-489f-a4fa-7d2e64fd36b6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.163500] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24eb4b10-9964-4511-a343-f8503390b2e7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.217843] env[62914]: DEBUG oslo_concurrency.lockutils [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.277617] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be6d774a-db08-4cde-a3ca-eee48f211e72 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.286351] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-858cde0c-a083-4a31-b72c-ff273b0b94de {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.321368] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62caf590-e9a1-4734-90ea-c58121889011 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.333284] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62420faa-8c8b-4bff-b7eb-667964421e65 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.348494] env[62914]: DEBUG nova.compute.provider_tree [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.353884] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352788, 'name': ReconfigVM_Task, 'duration_secs': 0.840601} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.354512] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 412c706a-31ce-4a33-914d-5762b02b4c63/412c706a-31ce-4a33-914d-5762b02b4c63.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.355159] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c726747e-4566-471f-a91c-0f30a9b97b38 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.362923] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 936.362923] env[62914]: value = "task-1352790" [ 936.362923] env[62914]: _type = "Task" [ 936.362923] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.378020] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352790, 'name': Rename_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.414829] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352789, 'name': CreateSnapshot_Task, 'duration_secs': 0.451561} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.415106] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Created Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 936.415860] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd4be52-726e-4117-863b-b41ab02e145f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.430029] env[62914]: DEBUG nova.compute.manager [req-1d889ff5-e9a8-4dd8-80ac-8c68f466c1d3 req-cf272165-b412-41ec-8cd2-3e2fdf86e358 service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Received event network-vif-plugged-d41f0fe0-d70c-4f8c-a306-e14dc0107565 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.430029] env[62914]: DEBUG oslo_concurrency.lockutils [req-1d889ff5-e9a8-4dd8-80ac-8c68f466c1d3 req-cf272165-b412-41ec-8cd2-3e2fdf86e358 service nova] Acquiring lock "2575bc0d-cb47-4dab-a219-f84254bda47d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.430215] env[62914]: DEBUG oslo_concurrency.lockutils [req-1d889ff5-e9a8-4dd8-80ac-8c68f466c1d3 req-cf272165-b412-41ec-8cd2-3e2fdf86e358 service nova] Lock "2575bc0d-cb47-4dab-a219-f84254bda47d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.430361] env[62914]: DEBUG oslo_concurrency.lockutils [req-1d889ff5-e9a8-4dd8-80ac-8c68f466c1d3 req-cf272165-b412-41ec-8cd2-3e2fdf86e358 service nova] Lock "2575bc0d-cb47-4dab-a219-f84254bda47d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.430572] env[62914]: DEBUG nova.compute.manager [req-1d889ff5-e9a8-4dd8-80ac-8c68f466c1d3 req-cf272165-b412-41ec-8cd2-3e2fdf86e358 service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] No waiting events found dispatching network-vif-plugged-d41f0fe0-d70c-4f8c-a306-e14dc0107565 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 936.430719] env[62914]: WARNING nova.compute.manager [req-1d889ff5-e9a8-4dd8-80ac-8c68f466c1d3 req-cf272165-b412-41ec-8cd2-3e2fdf86e358 service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Received unexpected event network-vif-plugged-d41f0fe0-d70c-4f8c-a306-e14dc0107565 for instance with vm_state building and task_state spawning. [ 936.533930] env[62914]: DEBUG nova.network.neutron [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Successfully updated port: d41f0fe0-d70c-4f8c-a306-e14dc0107565 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 936.855720] env[62914]: DEBUG nova.scheduler.client.report [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 936.874650] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352790, 'name': Rename_Task, 'duration_secs': 0.216583} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.875513] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 936.875826] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e7852ca3-a769-445b-bc22-665dced1c900 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.882811] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 936.882811] env[62914]: value = "task-1352791" [ 936.882811] env[62914]: _type = "Task" [ 936.882811] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.890883] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352791, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.934836] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Creating linked-clone VM from snapshot {{(pid=62914) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 936.935185] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8e8c4557-b926-41a1-9d73-bb1f344ccd55 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.946166] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 936.946166] env[62914]: value = "task-1352792" [ 936.946166] env[62914]: _type = "Task" [ 936.946166] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.959823] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352792, 'name': CloneVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.037681] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "refresh_cache-2575bc0d-cb47-4dab-a219-f84254bda47d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.037857] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquired lock "refresh_cache-2575bc0d-cb47-4dab-a219-f84254bda47d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.037988] env[62914]: DEBUG nova.network.neutron [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 937.361753] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.273s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.364623] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.720s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.366473] env[62914]: INFO nova.compute.claims [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 937.397407] env[62914]: DEBUG oslo_vmware.api [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352791, 'name': PowerOnVM_Task, 'duration_secs': 0.496707} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.397870] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 937.398223] env[62914]: INFO nova.compute.manager [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Took 8.91 seconds to spawn the instance on the hypervisor. [ 937.398529] env[62914]: DEBUG nova.compute.manager [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.399909] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb1ebf7-ccd4-410e-910d-b58f151d96c2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.415397] env[62914]: INFO nova.network.neutron [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updating port 57172af7-5ed6-46d9-9d7b-3ef100c530f7 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 937.457846] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352792, 'name': CloneVM_Task} progress is 94%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.595196] env[62914]: DEBUG nova.network.neutron [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 937.768219] env[62914]: DEBUG nova.network.neutron [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Updating instance_info_cache with network_info: [{"id": "d41f0fe0-d70c-4f8c-a306-e14dc0107565", "address": "fa:16:3e:b0:af:bb", "network": {"id": "184ac684-c4e6-4468-aee8-1625d96dc0b4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1513916205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3055903b6f4443bbeb9897e75e34c49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd41f0fe0-d7", "ovs_interfaceid": "d41f0fe0-d70c-4f8c-a306-e14dc0107565", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.928082] env[62914]: INFO nova.compute.manager [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Took 25.82 seconds to build instance. [ 937.958157] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352792, 'name': CloneVM_Task} progress is 95%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.269972] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Releasing lock "refresh_cache-2575bc0d-cb47-4dab-a219-f84254bda47d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.270347] env[62914]: DEBUG nova.compute.manager [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Instance network_info: |[{"id": "d41f0fe0-d70c-4f8c-a306-e14dc0107565", "address": "fa:16:3e:b0:af:bb", "network": {"id": "184ac684-c4e6-4468-aee8-1625d96dc0b4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1513916205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3055903b6f4443bbeb9897e75e34c49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd41f0fe0-d7", "ovs_interfaceid": "d41f0fe0-d70c-4f8c-a306-e14dc0107565", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 938.270866] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:af:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8459aaf-d6a8-46fb-ad14-464ac3104695', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd41f0fe0-d70c-4f8c-a306-e14dc0107565', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.278403] env[62914]: DEBUG oslo.service.loopingcall [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.278628] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 938.278859] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f7b5400-4507-4cfe-80c4-0f25d9a1a268 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.300334] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.300334] env[62914]: value = "task-1352793" [ 938.300334] env[62914]: _type = "Task" [ 938.300334] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.308713] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352793, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.430588] env[62914]: DEBUG oslo_concurrency.lockutils [None req-16370835-9e81-4747-af4f-5745085b4a31 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "412c706a-31ce-4a33-914d-5762b02b4c63" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.328s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.459668] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352792, 'name': CloneVM_Task} progress is 95%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.539773] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e517282-08ab-4af6-9076-5c9d26649a79 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.549778] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9d1f28-9267-4714-9df0-09471fbe562d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.589743] env[62914]: DEBUG nova.compute.manager [req-26dd3f70-e573-44b9-9677-6d29be96da8d req-49c8351a-4cf8-45b9-8252-70f9b7f96051 service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Received event network-changed-d41f0fe0-d70c-4f8c-a306-e14dc0107565 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.590050] env[62914]: DEBUG nova.compute.manager [req-26dd3f70-e573-44b9-9677-6d29be96da8d req-49c8351a-4cf8-45b9-8252-70f9b7f96051 service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Refreshing instance network info cache due to event network-changed-d41f0fe0-d70c-4f8c-a306-e14dc0107565. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 938.590128] env[62914]: DEBUG oslo_concurrency.lockutils [req-26dd3f70-e573-44b9-9677-6d29be96da8d req-49c8351a-4cf8-45b9-8252-70f9b7f96051 service nova] Acquiring lock "refresh_cache-2575bc0d-cb47-4dab-a219-f84254bda47d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.590250] env[62914]: DEBUG oslo_concurrency.lockutils [req-26dd3f70-e573-44b9-9677-6d29be96da8d req-49c8351a-4cf8-45b9-8252-70f9b7f96051 service nova] Acquired lock "refresh_cache-2575bc0d-cb47-4dab-a219-f84254bda47d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.590418] env[62914]: DEBUG nova.network.neutron [req-26dd3f70-e573-44b9-9677-6d29be96da8d req-49c8351a-4cf8-45b9-8252-70f9b7f96051 service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Refreshing network info cache for port d41f0fe0-d70c-4f8c-a306-e14dc0107565 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 938.592132] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e84ab1a-c9d7-43d1-bb43-288856cd0c52 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.601740] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad51f2f-197d-4a34-a835-3361beb39740 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.620035] env[62914]: DEBUG nova.compute.provider_tree [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.810717] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352793, 'name': CreateVM_Task, 'duration_secs': 0.48939} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.810717] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 938.811517] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.811697] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.812050] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.812435] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab8e022d-c75b-4f34-9c82-8b1dd9023e74 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.817976] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 938.817976] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a71d4c-792d-c3c9-c4a4-2b0247e2a046" [ 938.817976] env[62914]: _type = "Task" [ 938.817976] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.826293] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a71d4c-792d-c3c9-c4a4-2b0247e2a046, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.963043] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352792, 'name': CloneVM_Task, 'duration_secs': 1.67044} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.963340] env[62914]: INFO nova.virt.vmwareapi.vmops [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Created linked-clone VM from snapshot [ 938.964372] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7fac60-9819-4a52-b4f2-77a1ff0c9290 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.972890] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Uploading image b461ea20-4eed-46de-86fc-18c0bc86beca {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 939.003322] env[62914]: DEBUG oslo_vmware.rw_handles [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 939.003322] env[62914]: value = "vm-288244" [ 939.003322] env[62914]: _type = "VirtualMachine" [ 939.003322] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 939.003621] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f671d1a4-6ba5-4d40-b894-ab71f90f6b7a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.013528] env[62914]: DEBUG oslo_vmware.rw_handles [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lease: (returnval){ [ 939.013528] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]524921ef-18c4-38d9-f9dc-193cc023de6e" [ 939.013528] env[62914]: _type = "HttpNfcLease" [ 939.013528] env[62914]: } obtained for exporting VM: (result){ [ 939.013528] env[62914]: value = "vm-288244" [ 939.013528] env[62914]: _type = "VirtualMachine" [ 939.013528] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 939.013886] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the lease: (returnval){ [ 939.013886] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]524921ef-18c4-38d9-f9dc-193cc023de6e" [ 939.013886] env[62914]: _type = "HttpNfcLease" [ 939.013886] env[62914]: } to be ready. {{(pid=62914) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 939.022655] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 939.022655] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]524921ef-18c4-38d9-f9dc-193cc023de6e" [ 939.022655] env[62914]: _type = "HttpNfcLease" [ 939.022655] env[62914]: } is initializing. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 939.038865] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.039077] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquired lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.039259] env[62914]: DEBUG nova.network.neutron [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 939.123823] env[62914]: DEBUG nova.scheduler.client.report [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.294738] env[62914]: DEBUG nova.network.neutron [req-26dd3f70-e573-44b9-9677-6d29be96da8d req-49c8351a-4cf8-45b9-8252-70f9b7f96051 service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Updated VIF entry in instance network info cache for port d41f0fe0-d70c-4f8c-a306-e14dc0107565. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 939.295145] env[62914]: DEBUG nova.network.neutron [req-26dd3f70-e573-44b9-9677-6d29be96da8d req-49c8351a-4cf8-45b9-8252-70f9b7f96051 service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Updating instance_info_cache with network_info: [{"id": "d41f0fe0-d70c-4f8c-a306-e14dc0107565", "address": "fa:16:3e:b0:af:bb", "network": {"id": "184ac684-c4e6-4468-aee8-1625d96dc0b4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1513916205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3055903b6f4443bbeb9897e75e34c49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd41f0fe0-d7", "ovs_interfaceid": "d41f0fe0-d70c-4f8c-a306-e14dc0107565", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.329419] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a71d4c-792d-c3c9-c4a4-2b0247e2a046, 'name': SearchDatastore_Task, 'duration_secs': 0.037359} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.329757] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.329998] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.330264] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.330416] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.330602] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.330931] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0954d0c3-cf40-409c-8a92-421daa7812a1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.344154] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.344352] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 939.345095] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc34f1cb-2564-42de-bf4b-cde912da09c6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.350738] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 939.350738] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52e901ce-d6b9-c327-d9ec-0fe6f9261dcc" [ 939.350738] env[62914]: _type = "Task" [ 939.350738] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.359593] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52e901ce-d6b9-c327-d9ec-0fe6f9261dcc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.523061] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 939.523061] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]524921ef-18c4-38d9-f9dc-193cc023de6e" [ 939.523061] env[62914]: _type = "HttpNfcLease" [ 939.523061] env[62914]: } is ready. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 939.523492] env[62914]: DEBUG oslo_vmware.rw_handles [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 939.523492] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]524921ef-18c4-38d9-f9dc-193cc023de6e" [ 939.523492] env[62914]: _type = "HttpNfcLease" [ 939.523492] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 939.524119] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d01dc0b-3731-4bbf-a3e0-166a6f506258 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.531504] env[62914]: DEBUG oslo_vmware.rw_handles [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524bad89-2218-57c1-d7ba-0fcf9584d4d2/disk-0.vmdk from lease info. {{(pid=62914) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 939.531696] env[62914]: DEBUG oslo_vmware.rw_handles [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524bad89-2218-57c1-d7ba-0fcf9584d4d2/disk-0.vmdk for reading. {{(pid=62914) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 939.590017] env[62914]: DEBUG nova.compute.manager [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.592381] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3dcd759-7d65-4cdb-8833-a476bee77974 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.626701] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b70b4566-102e-4db0-8abd-0f16812a4095 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.631980] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.267s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.632197] env[62914]: DEBUG nova.compute.manager [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 939.636012] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.596s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.636260] env[62914]: DEBUG nova.objects.instance [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lazy-loading 'resources' on Instance uuid f616eac4-7c75-484e-bb64-5e227a08df8c {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.798024] env[62914]: DEBUG oslo_concurrency.lockutils [req-26dd3f70-e573-44b9-9677-6d29be96da8d req-49c8351a-4cf8-45b9-8252-70f9b7f96051 service nova] Releasing lock "refresh_cache-2575bc0d-cb47-4dab-a219-f84254bda47d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.835612] env[62914]: DEBUG nova.network.neutron [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updating instance_info_cache with network_info: [{"id": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "address": "fa:16:3e:91:13:35", "network": {"id": "c4298e11-ab16-4b37-be29-48fc406e74b0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-695272838-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.236", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "15315f8e77d749a4b2b0211c3b22b546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe38bb7e-8bcb-419d-868f-0dc105c69651", "external-id": "nsx-vlan-transportzone-432", "segmentation_id": 432, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57172af7-5e", "ovs_interfaceid": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.862890] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52e901ce-d6b9-c327-d9ec-0fe6f9261dcc, 'name': SearchDatastore_Task, 'duration_secs': 0.043238} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.863843] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4bbf8013-d1df-4801-a780-0bea78369068 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.869927] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 939.869927] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]525a9c64-4138-859a-3406-d92f7a716052" [ 939.869927] env[62914]: _type = "Task" [ 939.869927] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.878874] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]525a9c64-4138-859a-3406-d92f7a716052, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.107489] env[62914]: INFO nova.compute.manager [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] instance snapshotting [ 940.110714] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc3b72f8-5bc4-4336-9d81-3729c5e45689 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.136176] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b102fe11-a9e6-4cad-8e8f-35df0a41375a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.143111] env[62914]: DEBUG nova.compute.utils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 940.149046] env[62914]: DEBUG nova.compute.manager [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 940.149572] env[62914]: DEBUG nova.network.neutron [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 940.197286] env[62914]: DEBUG nova.policy [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '795b94761b144a1e99ed79838d9687ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9278818cebd842669b802a4e43e91774', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 940.322529] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f88d92f-4039-4470-a629-0f2cb64bf367 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.331848] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e529c38-5662-4691-8e48-f7288d25c2af {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.338864] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Releasing lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.368328] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b826c06-d5e0-413e-a218-3f12b06218ed {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.378216] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0665cc9-0ee3-4e29-a113-0d31196d6c4d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.384729] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]525a9c64-4138-859a-3406-d92f7a716052, 'name': SearchDatastore_Task, 'duration_secs': 0.011558} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.387452] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.388013] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 2575bc0d-cb47-4dab-a219-f84254bda47d/2575bc0d-cb47-4dab-a219-f84254bda47d.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 940.388365] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1ed3691-1948-408b-8267-195c97189cf7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.399932] env[62914]: DEBUG nova.compute.provider_tree [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.403597] env[62914]: DEBUG nova.virt.hardware [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='7b922e975a062415769cf5be1ba4fe6b',container_format='bare',created_at=2024-10-24T12:15:11Z,direct_url=,disk_format='vmdk',id=662656a7-d46d-4633-b79c-60fc2ed8c7dc,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-931130847-shelved',owner='15315f8e77d749a4b2b0211c3b22b546',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-24T12:15:25Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.404604] env[62914]: DEBUG nova.virt.hardware [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.404604] env[62914]: DEBUG nova.virt.hardware [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.404604] env[62914]: DEBUG nova.virt.hardware [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.404604] env[62914]: DEBUG nova.virt.hardware [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.404844] env[62914]: DEBUG nova.virt.hardware [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.405043] env[62914]: DEBUG nova.virt.hardware [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.405284] env[62914]: DEBUG nova.virt.hardware [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.405587] env[62914]: DEBUG nova.virt.hardware [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.405809] env[62914]: DEBUG nova.virt.hardware [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.406056] env[62914]: DEBUG nova.virt.hardware [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.407395] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a79fd0e-a9b8-4c50-86d1-5aec8cb36760 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.413485] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 940.413485] env[62914]: value = "task-1352795" [ 940.413485] env[62914]: _type = "Task" [ 940.413485] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.420952] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e480248-e850-4435-9bb9-f39f52d6fb34 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.430745] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352795, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.439520] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:13:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe38bb7e-8bcb-419d-868f-0dc105c69651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '57172af7-5ed6-46d9-9d7b-3ef100c530f7', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.448838] env[62914]: DEBUG oslo.service.loopingcall [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.448838] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 940.448838] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e27abc45-1002-46b7-a8d2-d48e2f0730eb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.472435] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.472435] env[62914]: value = "task-1352796" [ 940.472435] env[62914]: _type = "Task" [ 940.472435] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.476792] env[62914]: DEBUG nova.network.neutron [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Successfully created port: 0c5a4114-78d6-4319-9b49-f26a44441811 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 940.485282] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352796, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.599341] env[62914]: DEBUG nova.compute.manager [req-97151274-2233-4631-ae63-1fc47ce45ac6 req-b4cab934-3c54-4494-98dd-cd89fe3a7ac1 service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Received event network-vif-plugged-57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.599637] env[62914]: DEBUG oslo_concurrency.lockutils [req-97151274-2233-4631-ae63-1fc47ce45ac6 req-b4cab934-3c54-4494-98dd-cd89fe3a7ac1 service nova] Acquiring lock "c7cf71d7-b823-471f-949f-9ef2b539d972-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.599817] env[62914]: DEBUG oslo_concurrency.lockutils [req-97151274-2233-4631-ae63-1fc47ce45ac6 req-b4cab934-3c54-4494-98dd-cd89fe3a7ac1 service nova] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.599991] env[62914]: DEBUG oslo_concurrency.lockutils [req-97151274-2233-4631-ae63-1fc47ce45ac6 req-b4cab934-3c54-4494-98dd-cd89fe3a7ac1 service nova] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.600587] env[62914]: DEBUG nova.compute.manager [req-97151274-2233-4631-ae63-1fc47ce45ac6 req-b4cab934-3c54-4494-98dd-cd89fe3a7ac1 service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] No waiting events found dispatching network-vif-plugged-57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 940.600794] env[62914]: WARNING nova.compute.manager [req-97151274-2233-4631-ae63-1fc47ce45ac6 req-b4cab934-3c54-4494-98dd-cd89fe3a7ac1 service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Received unexpected event network-vif-plugged-57172af7-5ed6-46d9-9d7b-3ef100c530f7 for instance with vm_state shelved_offloaded and task_state spawning. [ 940.600959] env[62914]: DEBUG nova.compute.manager [req-97151274-2233-4631-ae63-1fc47ce45ac6 req-b4cab934-3c54-4494-98dd-cd89fe3a7ac1 service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Received event network-changed-57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.601144] env[62914]: DEBUG nova.compute.manager [req-97151274-2233-4631-ae63-1fc47ce45ac6 req-b4cab934-3c54-4494-98dd-cd89fe3a7ac1 service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Refreshing instance network info cache due to event network-changed-57172af7-5ed6-46d9-9d7b-3ef100c530f7. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 940.601335] env[62914]: DEBUG oslo_concurrency.lockutils [req-97151274-2233-4631-ae63-1fc47ce45ac6 req-b4cab934-3c54-4494-98dd-cd89fe3a7ac1 service nova] Acquiring lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.601548] env[62914]: DEBUG oslo_concurrency.lockutils [req-97151274-2233-4631-ae63-1fc47ce45ac6 req-b4cab934-3c54-4494-98dd-cd89fe3a7ac1 service nova] Acquired lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.601733] env[62914]: DEBUG nova.network.neutron [req-97151274-2233-4631-ae63-1fc47ce45ac6 req-b4cab934-3c54-4494-98dd-cd89fe3a7ac1 service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Refreshing network info cache for port 57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.655646] env[62914]: DEBUG nova.compute.manager [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 940.662751] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Creating Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 940.663262] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e03a8c6b-f870-469e-935d-9eb5dc846ac1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.675202] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 940.675202] env[62914]: value = "task-1352797" [ 940.675202] env[62914]: _type = "Task" [ 940.675202] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.687256] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352797, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.912418] env[62914]: DEBUG nova.scheduler.client.report [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.932467] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352795, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.984097] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352796, 'name': CreateVM_Task, 'duration_secs': 0.424307} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.984327] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 940.985068] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/662656a7-d46d-4633-b79c-60fc2ed8c7dc" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.985261] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquired lock "[datastore1] devstack-image-cache_base/662656a7-d46d-4633-b79c-60fc2ed8c7dc" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.985648] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/662656a7-d46d-4633-b79c-60fc2ed8c7dc" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 940.985925] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-123d9233-456c-4255-8092-49e482803396 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.991608] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 940.991608] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c4f064-559d-e641-0f1a-aecdf60a68c7" [ 940.991608] env[62914]: _type = "Task" [ 940.991608] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.000973] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c4f064-559d-e641-0f1a-aecdf60a68c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.188717] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352797, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.326893] env[62914]: DEBUG nova.network.neutron [req-97151274-2233-4631-ae63-1fc47ce45ac6 req-b4cab934-3c54-4494-98dd-cd89fe3a7ac1 service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updated VIF entry in instance network info cache for port 57172af7-5ed6-46d9-9d7b-3ef100c530f7. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 941.327122] env[62914]: DEBUG nova.network.neutron [req-97151274-2233-4631-ae63-1fc47ce45ac6 req-b4cab934-3c54-4494-98dd-cd89fe3a7ac1 service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updating instance_info_cache with network_info: [{"id": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "address": "fa:16:3e:91:13:35", "network": {"id": "c4298e11-ab16-4b37-be29-48fc406e74b0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-695272838-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.236", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "15315f8e77d749a4b2b0211c3b22b546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe38bb7e-8bcb-419d-868f-0dc105c69651", "external-id": "nsx-vlan-transportzone-432", "segmentation_id": 432, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57172af7-5e", "ovs_interfaceid": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.424617] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.788s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.430253] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.366s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.430487] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.430587] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62914) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 941.430895] env[62914]: DEBUG oslo_concurrency.lockutils [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.213s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.431159] env[62914]: DEBUG nova.objects.instance [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lazy-loading 'resources' on Instance uuid 1f7b1609-08e0-4439-aea0-b4a2ec808780 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.432346] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352795, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.592829} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.434694] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b22679f-2463-4b37-83e0-941d1e89006e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.436092] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 2575bc0d-cb47-4dab-a219-f84254bda47d/2575bc0d-cb47-4dab-a219-f84254bda47d.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 941.436322] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 941.436630] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-80c99466-2b48-4f43-836f-a61cc5aa8935 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.447998] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ba8e13-2f4c-4d82-b0cd-7dd8daa3b4ff {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.453277] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 941.453277] env[62914]: value = "task-1352798" [ 941.453277] env[62914]: _type = "Task" [ 941.453277] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.466150] env[62914]: INFO nova.scheduler.client.report [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Deleted allocations for instance f616eac4-7c75-484e-bb64-5e227a08df8c [ 941.470894] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce4eb04-3eb7-477b-a349-869cd15532b4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.477402] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352798, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.483072] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d4a087-77d2-40e5-abab-1883655eee95 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.517128] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180876MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62914) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 941.517322] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.529226] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Releasing lock "[datastore1] devstack-image-cache_base/662656a7-d46d-4633-b79c-60fc2ed8c7dc" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.529528] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Processing image 662656a7-d46d-4633-b79c-60fc2ed8c7dc {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.529780] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/662656a7-d46d-4633-b79c-60fc2ed8c7dc/662656a7-d46d-4633-b79c-60fc2ed8c7dc.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.529944] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquired lock "[datastore1] devstack-image-cache_base/662656a7-d46d-4633-b79c-60fc2ed8c7dc/662656a7-d46d-4633-b79c-60fc2ed8c7dc.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.530143] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.530409] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c27559fe-61db-42d2-91ca-6124c459f3c8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.539485] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.539689] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 941.540421] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9968966-a89d-431a-80c7-45fd69d1d370 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.547019] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 941.547019] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]521b9771-2c74-d86d-2726-c963a3e95ce3" [ 941.547019] env[62914]: _type = "Task" [ 941.547019] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.555386] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]521b9771-2c74-d86d-2726-c963a3e95ce3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.675567] env[62914]: DEBUG nova.compute.manager [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 941.688874] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352797, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.701623] env[62914]: DEBUG nova.virt.hardware [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.701931] env[62914]: DEBUG nova.virt.hardware [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.702125] env[62914]: DEBUG nova.virt.hardware [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.702319] env[62914]: DEBUG nova.virt.hardware [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.702526] env[62914]: DEBUG nova.virt.hardware [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.702772] env[62914]: DEBUG nova.virt.hardware [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.703046] env[62914]: DEBUG nova.virt.hardware [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.703243] env[62914]: DEBUG nova.virt.hardware [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.703425] env[62914]: DEBUG nova.virt.hardware [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.703593] env[62914]: DEBUG nova.virt.hardware [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.703805] env[62914]: DEBUG nova.virt.hardware [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.704737] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6306f4-66bf-4d53-9bab-7cd0cd1865e3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.713133] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f481e01-e9ec-4b33-a959-2e2cd1a1a061 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.829986] env[62914]: DEBUG oslo_concurrency.lockutils [req-97151274-2233-4631-ae63-1fc47ce45ac6 req-b4cab934-3c54-4494-98dd-cd89fe3a7ac1 service nova] Releasing lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.966945] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352798, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073252} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.967273] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.968086] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a1159c-9dd8-4789-97b8-abb88f4596ac {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.999492] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 2575bc0d-cb47-4dab-a219-f84254bda47d/2575bc0d-cb47-4dab-a219-f84254bda47d.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.000216] env[62914]: DEBUG oslo_concurrency.lockutils [None req-fb5499bb-bb0f-4298-bc76-1aaa3a5e0203 tempest-ServerDiskConfigTestJSON-1466581344 tempest-ServerDiskConfigTestJSON-1466581344-project-member] Lock "f616eac4-7c75-484e-bb64-5e227a08df8c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.890s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.004647] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc316fec-e9ea-486c-8c89-4d38d044e0a4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.031198] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 942.031198] env[62914]: value = "task-1352799" [ 942.031198] env[62914]: _type = "Task" [ 942.031198] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.045074] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352799, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.059598] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Preparing fetch location {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 942.059918] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Fetch image to [datastore1] OSTACK_IMG_91237e19-c479-44c2-b070-0a49f1a8a0f6/OSTACK_IMG_91237e19-c479-44c2-b070-0a49f1a8a0f6.vmdk {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 942.060116] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Downloading stream optimized image 662656a7-d46d-4633-b79c-60fc2ed8c7dc to [datastore1] OSTACK_IMG_91237e19-c479-44c2-b070-0a49f1a8a0f6/OSTACK_IMG_91237e19-c479-44c2-b070-0a49f1a8a0f6.vmdk on the data store datastore1 as vApp {{(pid=62914) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 942.060296] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Downloading image file data 662656a7-d46d-4633-b79c-60fc2ed8c7dc to the ESX as VM named 'OSTACK_IMG_91237e19-c479-44c2-b070-0a49f1a8a0f6' {{(pid=62914) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 942.127406] env[62914]: DEBUG nova.network.neutron [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Successfully updated port: 0c5a4114-78d6-4319-9b49-f26a44441811 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 942.151949] env[62914]: DEBUG oslo_vmware.rw_handles [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 942.151949] env[62914]: value = "resgroup-9" [ 942.151949] env[62914]: _type = "ResourcePool" [ 942.151949] env[62914]: }. {{(pid=62914) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 942.152482] env[62914]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-cb0d63ce-2191-4de1-bac1-ac04d1a39134 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.178596] env[62914]: DEBUG oslo_vmware.rw_handles [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lease: (returnval){ [ 942.178596] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]526950be-a68f-8bb9-99e1-a15f39c9e75d" [ 942.178596] env[62914]: _type = "HttpNfcLease" [ 942.178596] env[62914]: } obtained for vApp import into resource pool (val){ [ 942.178596] env[62914]: value = "resgroup-9" [ 942.178596] env[62914]: _type = "ResourcePool" [ 942.178596] env[62914]: }. {{(pid=62914) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 942.179068] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the lease: (returnval){ [ 942.179068] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]526950be-a68f-8bb9-99e1-a15f39c9e75d" [ 942.179068] env[62914]: _type = "HttpNfcLease" [ 942.179068] env[62914]: } to be ready. {{(pid=62914) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 942.189896] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e93cc702-1ce6-4134-80f6-78d3bd692304 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.197620] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 942.197620] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]526950be-a68f-8bb9-99e1-a15f39c9e75d" [ 942.197620] env[62914]: _type = "HttpNfcLease" [ 942.197620] env[62914]: } is initializing. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 942.197974] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352797, 'name': CreateSnapshot_Task, 'duration_secs': 1.055671} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.198773] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Created Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 942.199624] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658af38e-d356-47f0-af94-01f4dbe9c683 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.206205] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbea5119-59d0-4462-bf05-fae78ffeb146 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.246825] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-398c87e0-a3ca-439f-ab46-a63613075e36 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.256689] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c5dcfe5-fa01-4881-a994-993e169960fd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.272649] env[62914]: DEBUG nova.compute.provider_tree [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.544051] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352799, 'name': ReconfigVM_Task, 'duration_secs': 0.459773} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.544394] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 2575bc0d-cb47-4dab-a219-f84254bda47d/2575bc0d-cb47-4dab-a219-f84254bda47d.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.545048] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-82efd450-07c1-47be-837d-7e1d98335b32 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.552775] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 942.552775] env[62914]: value = "task-1352801" [ 942.552775] env[62914]: _type = "Task" [ 942.552775] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.562499] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352801, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.631673] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "refresh_cache-10c348bd-d85d-49b8-9b0f-216142f664d1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.631912] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "refresh_cache-10c348bd-d85d-49b8-9b0f-216142f664d1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.632309] env[62914]: DEBUG nova.network.neutron [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 942.634890] env[62914]: DEBUG nova.compute.manager [req-e7f44069-ff5b-4b25-a937-0539ef97f2da req-fafde7d5-622c-4200-a710-c1e37de7f260 service nova] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Received event network-vif-plugged-0c5a4114-78d6-4319-9b49-f26a44441811 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.635242] env[62914]: DEBUG oslo_concurrency.lockutils [req-e7f44069-ff5b-4b25-a937-0539ef97f2da req-fafde7d5-622c-4200-a710-c1e37de7f260 service nova] Acquiring lock "10c348bd-d85d-49b8-9b0f-216142f664d1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.635534] env[62914]: DEBUG oslo_concurrency.lockutils [req-e7f44069-ff5b-4b25-a937-0539ef97f2da req-fafde7d5-622c-4200-a710-c1e37de7f260 service nova] Lock "10c348bd-d85d-49b8-9b0f-216142f664d1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.635787] env[62914]: DEBUG oslo_concurrency.lockutils [req-e7f44069-ff5b-4b25-a937-0539ef97f2da req-fafde7d5-622c-4200-a710-c1e37de7f260 service nova] Lock "10c348bd-d85d-49b8-9b0f-216142f664d1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.636127] env[62914]: DEBUG nova.compute.manager [req-e7f44069-ff5b-4b25-a937-0539ef97f2da req-fafde7d5-622c-4200-a710-c1e37de7f260 service nova] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] No waiting events found dispatching network-vif-plugged-0c5a4114-78d6-4319-9b49-f26a44441811 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 942.636611] env[62914]: WARNING nova.compute.manager [req-e7f44069-ff5b-4b25-a937-0539ef97f2da req-fafde7d5-622c-4200-a710-c1e37de7f260 service nova] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Received unexpected event network-vif-plugged-0c5a4114-78d6-4319-9b49-f26a44441811 for instance with vm_state building and task_state spawning. [ 942.636611] env[62914]: DEBUG nova.compute.manager [req-e7f44069-ff5b-4b25-a937-0539ef97f2da req-fafde7d5-622c-4200-a710-c1e37de7f260 service nova] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Received event network-changed-0c5a4114-78d6-4319-9b49-f26a44441811 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.636904] env[62914]: DEBUG nova.compute.manager [req-e7f44069-ff5b-4b25-a937-0539ef97f2da req-fafde7d5-622c-4200-a710-c1e37de7f260 service nova] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Refreshing instance network info cache due to event network-changed-0c5a4114-78d6-4319-9b49-f26a44441811. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 942.637221] env[62914]: DEBUG oslo_concurrency.lockutils [req-e7f44069-ff5b-4b25-a937-0539ef97f2da req-fafde7d5-622c-4200-a710-c1e37de7f260 service nova] Acquiring lock "refresh_cache-10c348bd-d85d-49b8-9b0f-216142f664d1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.689273] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 942.689273] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]526950be-a68f-8bb9-99e1-a15f39c9e75d" [ 942.689273] env[62914]: _type = "HttpNfcLease" [ 942.689273] env[62914]: } is initializing. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 942.729738] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Creating linked-clone VM from snapshot {{(pid=62914) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 942.730120] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-faf4da5e-1861-4a9d-b146-cf0f8c39400c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.741241] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 942.741241] env[62914]: value = "task-1352802" [ 942.741241] env[62914]: _type = "Task" [ 942.741241] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.750717] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352802, 'name': CloneVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.776017] env[62914]: DEBUG nova.scheduler.client.report [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.063952] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352801, 'name': Rename_Task, 'duration_secs': 0.190677} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.064330] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 943.064613] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5e414c3-2ee5-4d74-b0c6-ef44e8690ae6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.072784] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 943.072784] env[62914]: value = "task-1352803" [ 943.072784] env[62914]: _type = "Task" [ 943.072784] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.089563] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352803, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.175423] env[62914]: DEBUG nova.network.neutron [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 943.188273] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 943.188273] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]526950be-a68f-8bb9-99e1-a15f39c9e75d" [ 943.188273] env[62914]: _type = "HttpNfcLease" [ 943.188273] env[62914]: } is ready. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 943.188611] env[62914]: DEBUG oslo_vmware.rw_handles [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 943.188611] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]526950be-a68f-8bb9-99e1-a15f39c9e75d" [ 943.188611] env[62914]: _type = "HttpNfcLease" [ 943.188611] env[62914]: }. {{(pid=62914) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 943.189423] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22dc262-df0c-47ac-8161-dec02d675ce4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.198522] env[62914]: DEBUG oslo_vmware.rw_handles [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52da210b-bd41-7721-8479-83ec57a6e2dd/disk-0.vmdk from lease info. {{(pid=62914) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 943.198770] env[62914]: DEBUG oslo_vmware.rw_handles [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52da210b-bd41-7721-8479-83ec57a6e2dd/disk-0.vmdk. {{(pid=62914) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 943.271436] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b6e17bc2-0387-4456-a621-27184c3aa551 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.279415] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352802, 'name': CloneVM_Task} progress is 94%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.281949] env[62914]: DEBUG oslo_concurrency.lockutils [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.851s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.287378] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.768s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.322113] env[62914]: INFO nova.scheduler.client.report [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Deleted allocations for instance 1f7b1609-08e0-4439-aea0-b4a2ec808780 [ 943.425874] env[62914]: DEBUG nova.network.neutron [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Updating instance_info_cache with network_info: [{"id": "0c5a4114-78d6-4319-9b49-f26a44441811", "address": "fa:16:3e:46:83:19", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c5a4114-78", "ovs_interfaceid": "0c5a4114-78d6-4319-9b49-f26a44441811", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.583920] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352803, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.778026] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352802, 'name': CloneVM_Task} progress is 94%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.831206] env[62914]: DEBUG oslo_concurrency.lockutils [None req-260c9858-fa06-4bfc-b9d0-f9416535c048 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "1f7b1609-08e0-4439-aea0-b4a2ec808780" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.501s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.932436] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "refresh_cache-10c348bd-d85d-49b8-9b0f-216142f664d1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.933137] env[62914]: DEBUG nova.compute.manager [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Instance network_info: |[{"id": "0c5a4114-78d6-4319-9b49-f26a44441811", "address": "fa:16:3e:46:83:19", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c5a4114-78", "ovs_interfaceid": "0c5a4114-78d6-4319-9b49-f26a44441811", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 943.933321] env[62914]: DEBUG oslo_concurrency.lockutils [req-e7f44069-ff5b-4b25-a937-0539ef97f2da req-fafde7d5-622c-4200-a710-c1e37de7f260 service nova] Acquired lock "refresh_cache-10c348bd-d85d-49b8-9b0f-216142f664d1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.933517] env[62914]: DEBUG nova.network.neutron [req-e7f44069-ff5b-4b25-a937-0539ef97f2da req-fafde7d5-622c-4200-a710-c1e37de7f260 service nova] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Refreshing network info cache for port 0c5a4114-78d6-4319-9b49-f26a44441811 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 943.935069] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:83:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '32faf59b-014c-4f1f-8331-40df95bf741f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c5a4114-78d6-4319-9b49-f26a44441811', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 943.944614] env[62914]: DEBUG oslo.service.loopingcall [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.950103] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 943.950711] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b4f93f6b-4791-491f-ace7-fbcf4992b511 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.982835] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 943.982835] env[62914]: value = "task-1352804" [ 943.982835] env[62914]: _type = "Task" [ 943.982835] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.999864] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352804, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.097349] env[62914]: DEBUG oslo_vmware.api [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352803, 'name': PowerOnVM_Task, 'duration_secs': 0.591963} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.097706] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 944.097932] env[62914]: INFO nova.compute.manager [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Took 7.98 seconds to spawn the instance on the hypervisor. [ 944.098143] env[62914]: DEBUG nova.compute.manager [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.098978] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5822d44c-aff3-4040-8d5b-95977e5ab8ff {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.280515] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352802, 'name': CloneVM_Task, 'duration_secs': 1.414244} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.281240] env[62914]: INFO nova.virt.vmwareapi.vmops [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Created linked-clone VM from snapshot [ 944.282719] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d133d57f-b2b4-4828-af42-bfab435bc355 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.296000] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Uploading image 67e90822-61d3-4074-a5de-33c47f66782e {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 944.332022] env[62914]: DEBUG oslo_vmware.rw_handles [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 944.332022] env[62914]: value = "vm-288249" [ 944.332022] env[62914]: _type = "VirtualMachine" [ 944.332022] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 944.332022] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-05060360-d144-4f09-bb6f-74d4dd146b5a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.339442] env[62914]: DEBUG oslo_vmware.rw_handles [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lease: (returnval){ [ 944.339442] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52bf141c-f812-693a-2869-4f8ee5c4a337" [ 944.339442] env[62914]: _type = "HttpNfcLease" [ 944.339442] env[62914]: } obtained for exporting VM: (result){ [ 944.339442] env[62914]: value = "vm-288249" [ 944.339442] env[62914]: _type = "VirtualMachine" [ 944.339442] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 944.339801] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the lease: (returnval){ [ 944.339801] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52bf141c-f812-693a-2869-4f8ee5c4a337" [ 944.339801] env[62914]: _type = "HttpNfcLease" [ 944.339801] env[62914]: } to be ready. {{(pid=62914) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 944.342340] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance ab036e1f-a62a-490a-818c-2e1256c781f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.342547] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance b8aebca4-2724-419e-aaf7-d180e299a459 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.345803] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 1e527707-758c-4006-af83-9c739b9645ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.345803] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 4bec4c88-bd1e-4cdf-9c8a-eee56088479b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.345803] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 13954007-b5ee-4254-b99e-5fbb2f6e0add actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.345803] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 412c706a-31ce-4a33-914d-5762b02b4c63 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.345803] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance c7cf71d7-b823-471f-949f-9ef2b539d972 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.346115] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 2575bc0d-cb47-4dab-a219-f84254bda47d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.346115] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 10c348bd-d85d-49b8-9b0f-216142f664d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 944.346202] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 944.346324] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 944.360119] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 944.360119] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52bf141c-f812-693a-2869-4f8ee5c4a337" [ 944.360119] env[62914]: _type = "HttpNfcLease" [ 944.360119] env[62914]: } is ready. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 944.362270] env[62914]: DEBUG oslo_vmware.rw_handles [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 944.362270] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52bf141c-f812-693a-2869-4f8ee5c4a337" [ 944.362270] env[62914]: _type = "HttpNfcLease" [ 944.362270] env[62914]: }. {{(pid=62914) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 944.362270] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f36f19-0c34-4fa2-aade-9e5ab86ecbd1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.369243] env[62914]: DEBUG nova.network.neutron [req-e7f44069-ff5b-4b25-a937-0539ef97f2da req-fafde7d5-622c-4200-a710-c1e37de7f260 service nova] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Updated VIF entry in instance network info cache for port 0c5a4114-78d6-4319-9b49-f26a44441811. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 944.369725] env[62914]: DEBUG nova.network.neutron [req-e7f44069-ff5b-4b25-a937-0539ef97f2da req-fafde7d5-622c-4200-a710-c1e37de7f260 service nova] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Updating instance_info_cache with network_info: [{"id": "0c5a4114-78d6-4319-9b49-f26a44441811", "address": "fa:16:3e:46:83:19", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c5a4114-78", "ovs_interfaceid": "0c5a4114-78d6-4319-9b49-f26a44441811", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.375396] env[62914]: DEBUG oslo_vmware.rw_handles [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f1f04a-6178-04bc-c77e-75f180d25546/disk-0.vmdk from lease info. {{(pid=62914) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 944.375602] env[62914]: DEBUG oslo_vmware.rw_handles [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f1f04a-6178-04bc-c77e-75f180d25546/disk-0.vmdk for reading. {{(pid=62914) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 944.479666] env[62914]: DEBUG oslo_concurrency.lockutils [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "b8aebca4-2724-419e-aaf7-d180e299a459" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.480054] env[62914]: DEBUG oslo_concurrency.lockutils [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "b8aebca4-2724-419e-aaf7-d180e299a459" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.480167] env[62914]: DEBUG oslo_concurrency.lockutils [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "b8aebca4-2724-419e-aaf7-d180e299a459-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.480727] env[62914]: DEBUG oslo_concurrency.lockutils [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "b8aebca4-2724-419e-aaf7-d180e299a459-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.480727] env[62914]: DEBUG oslo_concurrency.lockutils [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "b8aebca4-2724-419e-aaf7-d180e299a459-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.483954] env[62914]: INFO nova.compute.manager [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Terminating instance [ 944.490141] env[62914]: DEBUG nova.compute.manager [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 944.490276] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.491579] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894b5abb-f315-45b5-aadc-29b7fc03b823 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.505776] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.506117] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352804, 'name': CreateVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.506312] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f5cccc4-6bc8-42ee-a980-9b9f63ddf8d1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.515103] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cd71b694-336f-408f-ac80-9e30c20a4bb6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.522157] env[62914]: DEBUG oslo_vmware.api [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 944.522157] env[62914]: value = "task-1352806" [ 944.522157] env[62914]: _type = "Task" [ 944.522157] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.531478] env[62914]: DEBUG oslo_vmware.api [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352806, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.593313] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ece9e2b-623b-43ba-a4c1-bc1c9bb14b5b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.607589] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28adb2c1-456e-4ec9-a8e5-c81f6f718564 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.659373] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f252855b-bb65-4bd8-92d6-fa83ba1c4928 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.664958] env[62914]: INFO nova.compute.manager [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Took 20.83 seconds to build instance. [ 944.671916] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc381c9b-d965-4b85-a464-a5e295e7b4c3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.690263] env[62914]: DEBUG nova.compute.provider_tree [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.693915] env[62914]: DEBUG oslo_vmware.rw_handles [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Completed reading data from the image iterator. {{(pid=62914) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 944.694290] env[62914]: DEBUG oslo_vmware.rw_handles [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52da210b-bd41-7721-8479-83ec57a6e2dd/disk-0.vmdk. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 944.695866] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf5459b-5eb6-4018-b33f-30b87733bacd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.705441] env[62914]: DEBUG oslo_vmware.rw_handles [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52da210b-bd41-7721-8479-83ec57a6e2dd/disk-0.vmdk is in state: ready. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 944.705441] env[62914]: DEBUG oslo_vmware.rw_handles [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52da210b-bd41-7721-8479-83ec57a6e2dd/disk-0.vmdk. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 944.705441] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-1857520d-a6fa-452e-8e98-57214b26b085 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.880853] env[62914]: DEBUG oslo_concurrency.lockutils [req-e7f44069-ff5b-4b25-a937-0539ef97f2da req-fafde7d5-622c-4200-a710-c1e37de7f260 service nova] Releasing lock "refresh_cache-10c348bd-d85d-49b8-9b0f-216142f664d1" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.000627] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352804, 'name': CreateVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.030824] env[62914]: DEBUG oslo_vmware.api [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352806, 'name': PowerOffVM_Task, 'duration_secs': 0.279445} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.031299] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 945.031413] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 945.031791] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57dc1c0f-0f7a-4e5c-9a76-6e8e70bd3382 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.095433] env[62914]: DEBUG oslo_vmware.rw_handles [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52da210b-bd41-7721-8479-83ec57a6e2dd/disk-0.vmdk. {{(pid=62914) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 945.095433] env[62914]: INFO nova.virt.vmwareapi.images [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Downloaded image file data 662656a7-d46d-4633-b79c-60fc2ed8c7dc [ 945.095657] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01017f52-9604-424a-8c6a-1cd8c5a2ddc4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.116153] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d7eb4c2-3073-4f1f-b1dd-7dd13d0aabce {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.139783] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 945.140077] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 945.140472] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Deleting the datastore file [datastore2] b8aebca4-2724-419e-aaf7-d180e299a459 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.140618] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-576982ec-c72b-4adb-a74a-08a13b9b3e9e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.155770] env[62914]: INFO nova.virt.vmwareapi.images [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] The imported VM was unregistered [ 945.156946] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Caching image {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 945.157211] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Creating directory with path [datastore1] devstack-image-cache_base/662656a7-d46d-4633-b79c-60fc2ed8c7dc {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 945.157637] env[62914]: DEBUG oslo_vmware.api [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 945.157637] env[62914]: value = "task-1352809" [ 945.157637] env[62914]: _type = "Task" [ 945.157637] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.157873] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2bb2db7c-f971-42f5-949f-0b68113de223 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.167435] env[62914]: DEBUG oslo_concurrency.lockutils [None req-07dab908-7714-4fc5-8c8f-7a3c3309c0d7 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "2575bc0d-cb47-4dab-a219-f84254bda47d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.341s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.173223] env[62914]: DEBUG oslo_vmware.api [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352809, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.177973] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Created directory with path [datastore1] devstack-image-cache_base/662656a7-d46d-4633-b79c-60fc2ed8c7dc {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 945.178203] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_91237e19-c479-44c2-b070-0a49f1a8a0f6/OSTACK_IMG_91237e19-c479-44c2-b070-0a49f1a8a0f6.vmdk to [datastore1] devstack-image-cache_base/662656a7-d46d-4633-b79c-60fc2ed8c7dc/662656a7-d46d-4633-b79c-60fc2ed8c7dc.vmdk. {{(pid=62914) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 945.178512] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-ae27e716-fa45-4239-a29e-8373acd0dbf5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.188434] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 945.188434] env[62914]: value = "task-1352810" [ 945.188434] env[62914]: _type = "Task" [ 945.188434] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.200963] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.205387] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352810, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.495464] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352804, 'name': CreateVM_Task, 'duration_secs': 1.155421} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.495722] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 945.496619] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.496857] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.497273] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 945.497605] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ad457b7-ab69-4d1e-a3e1-2da94325d4ce {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.504258] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 945.504258] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]526d2e3f-fecc-0c24-7f50-2abc60e2fcf2" [ 945.504258] env[62914]: _type = "Task" [ 945.504258] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.515451] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]526d2e3f-fecc-0c24-7f50-2abc60e2fcf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.677737] env[62914]: DEBUG oslo_vmware.api [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352809, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.291484} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.678514] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 945.678792] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 945.679035] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 945.679532] env[62914]: INFO nova.compute.manager [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Took 1.19 seconds to destroy the instance on the hypervisor. [ 945.680069] env[62914]: DEBUG oslo.service.loopingcall [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.680445] env[62914]: DEBUG nova.compute.manager [-] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 945.680588] env[62914]: DEBUG nova.network.neutron [-] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 945.708314] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 945.708722] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.423s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.709156] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352810, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.859217] env[62914]: DEBUG nova.compute.manager [req-22a9de5d-7b60-488d-994f-d73f19b02b55 req-1cb3a3ac-d91c-4851-9176-49226850a739 service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Received event network-changed-d41f0fe0-d70c-4f8c-a306-e14dc0107565 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.863571] env[62914]: DEBUG nova.compute.manager [req-22a9de5d-7b60-488d-994f-d73f19b02b55 req-1cb3a3ac-d91c-4851-9176-49226850a739 service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Refreshing instance network info cache due to event network-changed-d41f0fe0-d70c-4f8c-a306-e14dc0107565. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 945.863571] env[62914]: DEBUG oslo_concurrency.lockutils [req-22a9de5d-7b60-488d-994f-d73f19b02b55 req-1cb3a3ac-d91c-4851-9176-49226850a739 service nova] Acquiring lock "refresh_cache-2575bc0d-cb47-4dab-a219-f84254bda47d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.863571] env[62914]: DEBUG oslo_concurrency.lockutils [req-22a9de5d-7b60-488d-994f-d73f19b02b55 req-1cb3a3ac-d91c-4851-9176-49226850a739 service nova] Acquired lock "refresh_cache-2575bc0d-cb47-4dab-a219-f84254bda47d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.863571] env[62914]: DEBUG nova.network.neutron [req-22a9de5d-7b60-488d-994f-d73f19b02b55 req-1cb3a3ac-d91c-4851-9176-49226850a739 service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Refreshing network info cache for port d41f0fe0-d70c-4f8c-a306-e14dc0107565 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.018373] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]526d2e3f-fecc-0c24-7f50-2abc60e2fcf2, 'name': SearchDatastore_Task, 'duration_secs': 0.074644} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.018768] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.020081] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.020081] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.020081] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.020081] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.020081] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a472eba-cdf6-433a-ae47-731a46c1817f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.038971] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.039556] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 946.040176] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f66ef420-8706-4407-a702-c7975e9cad4a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.049602] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 946.049602] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f01045-c076-bbe2-f312-db5b48e4b89a" [ 946.049602] env[62914]: _type = "Task" [ 946.049602] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.062283] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f01045-c076-bbe2-f312-db5b48e4b89a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.201305] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352810, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.496858] env[62914]: DEBUG oslo_vmware.rw_handles [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524bad89-2218-57c1-d7ba-0fcf9584d4d2/disk-0.vmdk. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 946.500355] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c69b80-ce1b-40cc-8ca5-26d47085fb83 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.507848] env[62914]: DEBUG oslo_vmware.rw_handles [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524bad89-2218-57c1-d7ba-0fcf9584d4d2/disk-0.vmdk is in state: ready. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 946.508114] env[62914]: ERROR oslo_vmware.rw_handles [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524bad89-2218-57c1-d7ba-0fcf9584d4d2/disk-0.vmdk due to incomplete transfer. [ 946.510854] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7a295b68-809c-440c-aadc-4022305bbb1a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.522335] env[62914]: DEBUG oslo_vmware.rw_handles [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524bad89-2218-57c1-d7ba-0fcf9584d4d2/disk-0.vmdk. {{(pid=62914) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 946.522335] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Uploaded image b461ea20-4eed-46de-86fc-18c0bc86beca to the Glance image server {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 946.525372] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Destroying the VM {{(pid=62914) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 946.525725] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4b7a0c10-8e9d-4e19-94ed-ccd74c52636c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.536721] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 946.536721] env[62914]: value = "task-1352811" [ 946.536721] env[62914]: _type = "Task" [ 946.536721] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.548906] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352811, 'name': Destroy_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.565782] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f01045-c076-bbe2-f312-db5b48e4b89a, 'name': SearchDatastore_Task, 'duration_secs': 0.090204} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.565782] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0de6ee0f-4331-4c9a-ae47-3c75ff676b1c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.566810] env[62914]: DEBUG nova.network.neutron [-] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.576570] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 946.576570] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]520f0d14-e779-f72c-f43c-4e259e7e275a" [ 946.576570] env[62914]: _type = "Task" [ 946.576570] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.584900] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]520f0d14-e779-f72c-f43c-4e259e7e275a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.627432] env[62914]: DEBUG nova.network.neutron [req-22a9de5d-7b60-488d-994f-d73f19b02b55 req-1cb3a3ac-d91c-4851-9176-49226850a739 service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Updated VIF entry in instance network info cache for port d41f0fe0-d70c-4f8c-a306-e14dc0107565. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 946.627852] env[62914]: DEBUG nova.network.neutron [req-22a9de5d-7b60-488d-994f-d73f19b02b55 req-1cb3a3ac-d91c-4851-9176-49226850a739 service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Updating instance_info_cache with network_info: [{"id": "d41f0fe0-d70c-4f8c-a306-e14dc0107565", "address": "fa:16:3e:b0:af:bb", "network": {"id": "184ac684-c4e6-4468-aee8-1625d96dc0b4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1513916205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c3055903b6f4443bbeb9897e75e34c49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd41f0fe0-d7", "ovs_interfaceid": "d41f0fe0-d70c-4f8c-a306-e14dc0107565", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.704350] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352810, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.052214] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352811, 'name': Destroy_Task, 'duration_secs': 0.420079} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.053184] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Destroyed the VM [ 947.054255] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Deleting Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 947.054636] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cad6e7b7-f2bd-44e2-9021-4acca426968a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.065691] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 947.065691] env[62914]: value = "task-1352812" [ 947.065691] env[62914]: _type = "Task" [ 947.065691] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.072833] env[62914]: INFO nova.compute.manager [-] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Took 1.39 seconds to deallocate network for instance. [ 947.089727] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352812, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.099072] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]520f0d14-e779-f72c-f43c-4e259e7e275a, 'name': SearchDatastore_Task, 'duration_secs': 0.087166} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.099483] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.099767] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 10c348bd-d85d-49b8-9b0f-216142f664d1/10c348bd-d85d-49b8-9b0f-216142f664d1.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 947.100096] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4687e766-eb80-4bf1-a1ee-63652e4d287a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.109657] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 947.109657] env[62914]: value = "task-1352813" [ 947.109657] env[62914]: _type = "Task" [ 947.109657] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.120950] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352813, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.130634] env[62914]: DEBUG oslo_concurrency.lockutils [req-22a9de5d-7b60-488d-994f-d73f19b02b55 req-1cb3a3ac-d91c-4851-9176-49226850a739 service nova] Releasing lock "refresh_cache-2575bc0d-cb47-4dab-a219-f84254bda47d" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.202921] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352810, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.580199] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352812, 'name': RemoveSnapshot_Task, 'duration_secs': 0.440594} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.580491] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Deleted Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 947.580758] env[62914]: DEBUG nova.compute.manager [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.582913] env[62914]: DEBUG oslo_concurrency.lockutils [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.584700] env[62914]: DEBUG oslo_concurrency.lockutils [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.584700] env[62914]: DEBUG nova.objects.instance [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lazy-loading 'resources' on Instance uuid b8aebca4-2724-419e-aaf7-d180e299a459 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.585105] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d75207f9-2a22-4fbb-9091-acc19eaefd0f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.623410] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352813, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.703407] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352810, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.881211] env[62914]: DEBUG nova.compute.manager [req-9373a51d-20f9-4077-8662-839928bf8f7a req-8d54aafa-1766-4bf7-ba74-0f96389a80e9 service nova] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Received event network-vif-deleted-ad5c411a-d657-4b8e-8cb7-1ef6d90a0bc2 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.106111] env[62914]: INFO nova.compute.manager [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Shelve offloading [ 948.110168] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 948.110595] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2eaddde5-b24b-468a-b12e-636218cf957f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.126304] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352813, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.127916] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 948.127916] env[62914]: value = "task-1352814" [ 948.127916] env[62914]: _type = "Task" [ 948.127916] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.143895] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] VM already powered off {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 948.144219] env[62914]: DEBUG nova.compute.manager [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.146399] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9fd9cd-c5f4-4aba-8370-5d0efcd1e1fe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.156555] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "refresh_cache-13954007-b5ee-4254-b99e-5fbb2f6e0add" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.156743] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "refresh_cache-13954007-b5ee-4254-b99e-5fbb2f6e0add" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.157234] env[62914]: DEBUG nova.network.neutron [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 948.205698] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352810, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.69004} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.208846] env[62914]: INFO nova.virt.vmwareapi.ds_util [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_91237e19-c479-44c2-b070-0a49f1a8a0f6/OSTACK_IMG_91237e19-c479-44c2-b070-0a49f1a8a0f6.vmdk to [datastore1] devstack-image-cache_base/662656a7-d46d-4633-b79c-60fc2ed8c7dc/662656a7-d46d-4633-b79c-60fc2ed8c7dc.vmdk. [ 948.209080] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Cleaning up location [datastore1] OSTACK_IMG_91237e19-c479-44c2-b070-0a49f1a8a0f6 {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 948.209259] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_91237e19-c479-44c2-b070-0a49f1a8a0f6 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 948.209834] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e2222b9-1ab2-4e5e-b98b-609703049c1c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.220075] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 948.220075] env[62914]: value = "task-1352815" [ 948.220075] env[62914]: _type = "Task" [ 948.220075] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.236095] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352815, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.265757] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f57c35d-9ec7-4d72-bc6d-f76a0f736f16 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.275390] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2fea318-9b17-49a4-af93-3345a751bd6a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.308086] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3feb07ed-c3ff-4042-9c14-68454f7c3f79 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.317572] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dade126-465e-4877-b8fa-4c304be31ad8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.333902] env[62914]: DEBUG nova.compute.provider_tree [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.625149] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352813, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.154315} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.625482] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 10c348bd-d85d-49b8-9b0f-216142f664d1/10c348bd-d85d-49b8-9b0f-216142f664d1.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 948.625717] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.626041] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4ac02306-3020-43b7-adaa-61a0fb69fe1f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.634588] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 948.634588] env[62914]: value = "task-1352816" [ 948.634588] env[62914]: _type = "Task" [ 948.634588] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.649831] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352816, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.731677] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352815, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.061886} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.731677] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 948.732083] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Releasing lock "[datastore1] devstack-image-cache_base/662656a7-d46d-4633-b79c-60fc2ed8c7dc/662656a7-d46d-4633-b79c-60fc2ed8c7dc.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.732083] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/662656a7-d46d-4633-b79c-60fc2ed8c7dc/662656a7-d46d-4633-b79c-60fc2ed8c7dc.vmdk to [datastore1] c7cf71d7-b823-471f-949f-9ef2b539d972/c7cf71d7-b823-471f-949f-9ef2b539d972.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 948.733158] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d018c040-2457-4828-bdc7-8ca17590e710 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.742026] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 948.742026] env[62914]: value = "task-1352817" [ 948.742026] env[62914]: _type = "Task" [ 948.742026] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.751233] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352817, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.836174] env[62914]: DEBUG nova.scheduler.client.report [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.915915] env[62914]: DEBUG nova.network.neutron [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Updating instance_info_cache with network_info: [{"id": "4e0349f9-75d4-4048-a903-03f627627c22", "address": "fa:16:3e:d0:25:1d", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e0349f9-75", "ovs_interfaceid": "4e0349f9-75d4-4048-a903-03f627627c22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.145783] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352816, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07043} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.146159] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 949.147657] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a143fe-bcb5-4fc1-8f66-13af54cc1c74 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.173479] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 10c348bd-d85d-49b8-9b0f-216142f664d1/10c348bd-d85d-49b8-9b0f-216142f664d1.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.173721] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02b9ef8a-c9fa-4fac-b6b1-3e25ec8f6cd0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.196571] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 949.196571] env[62914]: value = "task-1352818" [ 949.196571] env[62914]: _type = "Task" [ 949.196571] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.210729] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352818, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.253207] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352817, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.341547] env[62914]: DEBUG oslo_concurrency.lockutils [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.758s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.360887] env[62914]: INFO nova.scheduler.client.report [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Deleted allocations for instance b8aebca4-2724-419e-aaf7-d180e299a459 [ 949.418136] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "refresh_cache-13954007-b5ee-4254-b99e-5fbb2f6e0add" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.713314] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352818, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.733378] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 949.734778] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95a2a2b-38c3-4cf7-9fee-503958b4f9c6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.745913] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 949.749825] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10c5e38b-4536-4474-b1d8-c73efc08459f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.757234] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.757671] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.763893] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352817, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.869281] env[62914]: DEBUG oslo_concurrency.lockutils [None req-09b32414-e2ab-453a-b1b2-a9939efee208 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "b8aebca4-2724-419e-aaf7-d180e299a459" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.389s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.912007] env[62914]: DEBUG nova.compute.manager [req-1b806704-b1b5-4370-8b70-d6bce71ab9d2 req-02e8467a-792a-4763-b82b-3c491268e5ac service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Received event network-vif-unplugged-4e0349f9-75d4-4048-a903-03f627627c22 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.912575] env[62914]: DEBUG oslo_concurrency.lockutils [req-1b806704-b1b5-4370-8b70-d6bce71ab9d2 req-02e8467a-792a-4763-b82b-3c491268e5ac service nova] Acquiring lock "13954007-b5ee-4254-b99e-5fbb2f6e0add-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.912575] env[62914]: DEBUG oslo_concurrency.lockutils [req-1b806704-b1b5-4370-8b70-d6bce71ab9d2 req-02e8467a-792a-4763-b82b-3c491268e5ac service nova] Lock "13954007-b5ee-4254-b99e-5fbb2f6e0add-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.912949] env[62914]: DEBUG oslo_concurrency.lockutils [req-1b806704-b1b5-4370-8b70-d6bce71ab9d2 req-02e8467a-792a-4763-b82b-3c491268e5ac service nova] Lock "13954007-b5ee-4254-b99e-5fbb2f6e0add-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.912949] env[62914]: DEBUG nova.compute.manager [req-1b806704-b1b5-4370-8b70-d6bce71ab9d2 req-02e8467a-792a-4763-b82b-3c491268e5ac service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] No waiting events found dispatching network-vif-unplugged-4e0349f9-75d4-4048-a903-03f627627c22 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 949.913101] env[62914]: WARNING nova.compute.manager [req-1b806704-b1b5-4370-8b70-d6bce71ab9d2 req-02e8467a-792a-4763-b82b-3c491268e5ac service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Received unexpected event network-vif-unplugged-4e0349f9-75d4-4048-a903-03f627627c22 for instance with vm_state shelved and task_state shelving_offloading. [ 949.913192] env[62914]: DEBUG nova.compute.manager [req-1b806704-b1b5-4370-8b70-d6bce71ab9d2 req-02e8467a-792a-4763-b82b-3c491268e5ac service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Received event network-changed-4e0349f9-75d4-4048-a903-03f627627c22 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.913436] env[62914]: DEBUG nova.compute.manager [req-1b806704-b1b5-4370-8b70-d6bce71ab9d2 req-02e8467a-792a-4763-b82b-3c491268e5ac service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Refreshing instance network info cache due to event network-changed-4e0349f9-75d4-4048-a903-03f627627c22. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 949.913958] env[62914]: DEBUG oslo_concurrency.lockutils [req-1b806704-b1b5-4370-8b70-d6bce71ab9d2 req-02e8467a-792a-4763-b82b-3c491268e5ac service nova] Acquiring lock "refresh_cache-13954007-b5ee-4254-b99e-5fbb2f6e0add" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.913958] env[62914]: DEBUG oslo_concurrency.lockutils [req-1b806704-b1b5-4370-8b70-d6bce71ab9d2 req-02e8467a-792a-4763-b82b-3c491268e5ac service nova] Acquired lock "refresh_cache-13954007-b5ee-4254-b99e-5fbb2f6e0add" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.913958] env[62914]: DEBUG nova.network.neutron [req-1b806704-b1b5-4370-8b70-d6bce71ab9d2 req-02e8467a-792a-4763-b82b-3c491268e5ac service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Refreshing network info cache for port 4e0349f9-75d4-4048-a903-03f627627c22 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 950.145577] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 950.145939] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 950.146218] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleting the datastore file [datastore2] 13954007-b5ee-4254-b99e-5fbb2f6e0add {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.146625] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7e6654e1-e81b-44af-85c7-0e0fa87b0c24 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.156638] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 950.156638] env[62914]: value = "task-1352820" [ 950.156638] env[62914]: _type = "Task" [ 950.156638] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.167146] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352820, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.210765] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352818, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.257789] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352817, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.264709] env[62914]: DEBUG nova.compute.manager [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 950.281339] env[62914]: DEBUG oslo_concurrency.lockutils [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "ab036e1f-a62a-490a-818c-2e1256c781f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.281589] env[62914]: DEBUG oslo_concurrency.lockutils [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "ab036e1f-a62a-490a-818c-2e1256c781f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.281690] env[62914]: DEBUG oslo_concurrency.lockutils [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "ab036e1f-a62a-490a-818c-2e1256c781f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.282292] env[62914]: DEBUG oslo_concurrency.lockutils [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "ab036e1f-a62a-490a-818c-2e1256c781f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.282424] env[62914]: DEBUG oslo_concurrency.lockutils [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "ab036e1f-a62a-490a-818c-2e1256c781f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.285307] env[62914]: INFO nova.compute.manager [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Terminating instance [ 950.287210] env[62914]: DEBUG nova.compute.manager [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 950.288061] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 950.288532] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21464f4b-0588-4d2f-bfdd-4fd2becf2898 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.299394] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 950.299767] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a611d51a-2fe4-4457-89dc-c8866d456e17 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.312323] env[62914]: DEBUG oslo_vmware.api [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 950.312323] env[62914]: value = "task-1352821" [ 950.312323] env[62914]: _type = "Task" [ 950.312323] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.328445] env[62914]: DEBUG oslo_vmware.api [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352821, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.641979] env[62914]: DEBUG nova.network.neutron [req-1b806704-b1b5-4370-8b70-d6bce71ab9d2 req-02e8467a-792a-4763-b82b-3c491268e5ac service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Updated VIF entry in instance network info cache for port 4e0349f9-75d4-4048-a903-03f627627c22. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.642566] env[62914]: DEBUG nova.network.neutron [req-1b806704-b1b5-4370-8b70-d6bce71ab9d2 req-02e8467a-792a-4763-b82b-3c491268e5ac service nova] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Updating instance_info_cache with network_info: [{"id": "4e0349f9-75d4-4048-a903-03f627627c22", "address": "fa:16:3e:d0:25:1d", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": null, "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap4e0349f9-75", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.671544] env[62914]: DEBUG oslo_vmware.api [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352820, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.438396} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.672271] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.672530] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 950.672718] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 950.698435] env[62914]: INFO nova.scheduler.client.report [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleted allocations for instance 13954007-b5ee-4254-b99e-5fbb2f6e0add [ 950.717835] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352818, 'name': ReconfigVM_Task, 'duration_secs': 1.13228} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.718255] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 10c348bd-d85d-49b8-9b0f-216142f664d1/10c348bd-d85d-49b8-9b0f-216142f664d1.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.718977] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc1e6fa0-612b-4ca7-88b8-425a23043ce6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.729923] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 950.729923] env[62914]: value = "task-1352822" [ 950.729923] env[62914]: _type = "Task" [ 950.729923] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.741770] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352822, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.758539] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352817, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.788483] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.788795] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.790495] env[62914]: INFO nova.compute.claims [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.826264] env[62914]: DEBUG oslo_vmware.api [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352821, 'name': PowerOffVM_Task, 'duration_secs': 0.374513} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.826737] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 950.826997] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 950.827350] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-39d00ae0-ebf9-4fbb-8ee3-88ebfe6f0bd2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.920556] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 950.920879] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 950.921129] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Deleting the datastore file [datastore2] ab036e1f-a62a-490a-818c-2e1256c781f6 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.921432] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a7be6991-1eb4-4009-a281-a85f98876b2b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.931792] env[62914]: DEBUG oslo_vmware.api [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for the task: (returnval){ [ 950.931792] env[62914]: value = "task-1352824" [ 950.931792] env[62914]: _type = "Task" [ 950.931792] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.944853] env[62914]: DEBUG oslo_vmware.api [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352824, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.146211] env[62914]: DEBUG oslo_concurrency.lockutils [req-1b806704-b1b5-4370-8b70-d6bce71ab9d2 req-02e8467a-792a-4763-b82b-3c491268e5ac service nova] Releasing lock "refresh_cache-13954007-b5ee-4254-b99e-5fbb2f6e0add" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.203920] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.241742] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352822, 'name': Rename_Task, 'duration_secs': 0.4656} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.242190] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 951.242539] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3cfb8373-eae7-41ac-a306-409a385f050f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.254777] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 951.254777] env[62914]: value = "task-1352825" [ 951.254777] env[62914]: _type = "Task" [ 951.254777] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.261747] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352817, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.467749} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.262451] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/662656a7-d46d-4633-b79c-60fc2ed8c7dc/662656a7-d46d-4633-b79c-60fc2ed8c7dc.vmdk to [datastore1] c7cf71d7-b823-471f-949f-9ef2b539d972/c7cf71d7-b823-471f-949f-9ef2b539d972.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 951.263284] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3590b1-6f2b-4a96-b816-c245a497d0a1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.268764] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352825, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.288832] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] c7cf71d7-b823-471f-949f-9ef2b539d972/c7cf71d7-b823-471f-949f-9ef2b539d972.vmdk or device None with type streamOptimized {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 951.289199] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-575c956f-43dc-4e43-bbc4-73a75ad21b80 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.316731] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 951.316731] env[62914]: value = "task-1352826" [ 951.316731] env[62914]: _type = "Task" [ 951.316731] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.327801] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352826, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.445180] env[62914]: DEBUG oslo_vmware.api [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Task: {'id': task-1352824, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.319652} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.445543] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.445798] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 951.446058] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 951.446252] env[62914]: INFO nova.compute.manager [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Took 1.16 seconds to destroy the instance on the hypervisor. [ 951.446553] env[62914]: DEBUG oslo.service.loopingcall [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.446798] env[62914]: DEBUG nova.compute.manager [-] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 951.446960] env[62914]: DEBUG nova.network.neutron [-] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 951.765502] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352825, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.828102] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352826, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.876778] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ae834d04-897e-4198-a7f3-85fc110128c1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "13954007-b5ee-4254-b99e-5fbb2f6e0add" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.928711] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb37a6e-36ac-4fe7-9006-3eb41390bdc6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.937339] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a660817-1752-4359-aef0-f52df09235ea {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.943488] env[62914]: DEBUG nova.compute.manager [req-f2947379-023e-469e-8bdc-2cd53840d5c8 req-4f0f3e5b-e09a-4a87-919d-33ab88571730 service nova] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Received event network-vif-deleted-2282e6e9-7ccf-474e-9fd5-6738ed525721 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.943712] env[62914]: INFO nova.compute.manager [req-f2947379-023e-469e-8bdc-2cd53840d5c8 req-4f0f3e5b-e09a-4a87-919d-33ab88571730 service nova] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Neutron deleted interface 2282e6e9-7ccf-474e-9fd5-6738ed525721; detaching it from the instance and deleting it from the info cache [ 951.943907] env[62914]: DEBUG nova.network.neutron [req-f2947379-023e-469e-8bdc-2cd53840d5c8 req-4f0f3e5b-e09a-4a87-919d-33ab88571730 service nova] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.977112] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10604a29-2fbd-4d76-bb5e-1d4c57d3d872 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.987301] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36103dc1-a49d-4bef-bc77-295f4a0d6de1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.004023] env[62914]: DEBUG nova.compute.provider_tree [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.178698] env[62914]: DEBUG nova.network.neutron [-] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.268784] env[62914]: DEBUG oslo_vmware.api [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352825, 'name': PowerOnVM_Task, 'duration_secs': 0.952905} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.269166] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 952.269443] env[62914]: INFO nova.compute.manager [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Took 10.59 seconds to spawn the instance on the hypervisor. [ 952.269709] env[62914]: DEBUG nova.compute.manager [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 952.270856] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-629aca03-8359-4d64-a395-bdacf37853e2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.329152] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352826, 'name': ReconfigVM_Task, 'duration_secs': 0.831379} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.329467] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Reconfigured VM instance instance-00000048 to attach disk [datastore1] c7cf71d7-b823-471f-949f-9ef2b539d972/c7cf71d7-b823-471f-949f-9ef2b539d972.vmdk or device None with type streamOptimized {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 952.330248] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f42810f-90c8-4547-9108-26b8314ed390 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.339396] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 952.339396] env[62914]: value = "task-1352827" [ 952.339396] env[62914]: _type = "Task" [ 952.339396] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.348666] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352827, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.447157] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ed620e1-0a0c-4e95-b0cb-5e47eea25c4c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.457918] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1627375-8bd3-4edd-b871-5824934ae35f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.496012] env[62914]: DEBUG nova.compute.manager [req-f2947379-023e-469e-8bdc-2cd53840d5c8 req-4f0f3e5b-e09a-4a87-919d-33ab88571730 service nova] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Detach interface failed, port_id=2282e6e9-7ccf-474e-9fd5-6738ed525721, reason: Instance ab036e1f-a62a-490a-818c-2e1256c781f6 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 952.507412] env[62914]: DEBUG nova.scheduler.client.report [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.682581] env[62914]: INFO nova.compute.manager [-] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Took 1.24 seconds to deallocate network for instance. [ 952.791949] env[62914]: INFO nova.compute.manager [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Took 21.16 seconds to build instance. [ 952.850052] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352827, 'name': Rename_Task, 'duration_secs': 0.231999} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.850401] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.850751] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b64e47f8-7aee-44dc-8c73-5b3d5a4c1760 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.859791] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 952.859791] env[62914]: value = "task-1352828" [ 952.859791] env[62914]: _type = "Task" [ 952.859791] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.870114] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352828, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.012987] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.224s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.013789] env[62914]: DEBUG nova.compute.manager [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 953.017667] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.814s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.017972] env[62914]: DEBUG nova.objects.instance [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lazy-loading 'resources' on Instance uuid 13954007-b5ee-4254-b99e-5fbb2f6e0add {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.189277] env[62914]: DEBUG oslo_concurrency.lockutils [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.275803] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "10c348bd-d85d-49b8-9b0f-216142f664d1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.294656] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c95663e7-0c79-49df-bed8-5f45dc67c85e tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "10c348bd-d85d-49b8-9b0f-216142f664d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.689s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.295603] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "10c348bd-d85d-49b8-9b0f-216142f664d1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.020s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.295920] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "10c348bd-d85d-49b8-9b0f-216142f664d1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.296230] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "10c348bd-d85d-49b8-9b0f-216142f664d1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.296469] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "10c348bd-d85d-49b8-9b0f-216142f664d1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.299056] env[62914]: INFO nova.compute.manager [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Terminating instance [ 953.301262] env[62914]: DEBUG nova.compute.manager [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 953.301503] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 953.302411] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846ff419-98e7-4a55-a487-0d1de15fdb42 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.311635] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.311951] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9bef58c3-8be9-465d-8077-3298c67105a5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.319302] env[62914]: DEBUG oslo_vmware.api [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 953.319302] env[62914]: value = "task-1352829" [ 953.319302] env[62914]: _type = "Task" [ 953.319302] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.328967] env[62914]: DEBUG oslo_vmware.api [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352829, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.371234] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352828, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.522483] env[62914]: DEBUG nova.compute.utils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 953.525151] env[62914]: DEBUG nova.objects.instance [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lazy-loading 'numa_topology' on Instance uuid 13954007-b5ee-4254-b99e-5fbb2f6e0add {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.529037] env[62914]: DEBUG nova.compute.manager [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 953.529037] env[62914]: DEBUG nova.network.neutron [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 953.640568] env[62914]: DEBUG nova.policy [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '45afc180d8e441c78b16c28f6c18291b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bb65abdffe44acdac0cdb8701e3e27b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 953.833679] env[62914]: DEBUG oslo_vmware.api [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352829, 'name': PowerOffVM_Task, 'duration_secs': 0.252723} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.833968] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.834162] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 953.834423] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf4bb909-69cf-4631-9909-50dc88e0596c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.872835] env[62914]: DEBUG oslo_vmware.api [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352828, 'name': PowerOnVM_Task, 'duration_secs': 0.751437} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.873142] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 953.918719] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 953.919382] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 953.919723] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleting the datastore file [datastore1] 10c348bd-d85d-49b8-9b0f-216142f664d1 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.920238] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8c4f8369-3f96-43fb-9a9c-3f0b1434121d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.928858] env[62914]: DEBUG oslo_vmware.api [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 953.928858] env[62914]: value = "task-1352831" [ 953.928858] env[62914]: _type = "Task" [ 953.928858] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.938779] env[62914]: DEBUG oslo_vmware.api [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352831, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.990909] env[62914]: DEBUG nova.compute.manager [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.993285] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7f84e8-9281-4b41-b1db-8c139a9a01a5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.012419] env[62914]: DEBUG nova.network.neutron [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Successfully created port: e02166ec-c4c0-44a5-a9d6-5927b987648d {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 954.031614] env[62914]: DEBUG nova.objects.base [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Object Instance<13954007-b5ee-4254-b99e-5fbb2f6e0add> lazy-loaded attributes: resources,numa_topology {{(pid=62914) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 954.035226] env[62914]: DEBUG nova.compute.manager [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 954.080322] env[62914]: DEBUG oslo_vmware.rw_handles [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f1f04a-6178-04bc-c77e-75f180d25546/disk-0.vmdk. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 954.081460] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-124d69a2-0e1f-485a-b929-4aa2a5e4a567 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.088492] env[62914]: DEBUG oslo_vmware.rw_handles [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f1f04a-6178-04bc-c77e-75f180d25546/disk-0.vmdk is in state: ready. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 954.088492] env[62914]: ERROR oslo_vmware.rw_handles [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f1f04a-6178-04bc-c77e-75f180d25546/disk-0.vmdk due to incomplete transfer. [ 954.088728] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f7e4716d-ef1e-430f-9c3c-9ead05f168c2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.097278] env[62914]: DEBUG oslo_vmware.rw_handles [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f1f04a-6178-04bc-c77e-75f180d25546/disk-0.vmdk. {{(pid=62914) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 954.097520] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Uploaded image 67e90822-61d3-4074-a5de-33c47f66782e to the Glance image server {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 954.099312] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Destroying the VM {{(pid=62914) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 954.102688] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2b4d831a-06dd-4947-9509-15705d149744 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.110710] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 954.110710] env[62914]: value = "task-1352832" [ 954.110710] env[62914]: _type = "Task" [ 954.110710] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.119522] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352832, 'name': Destroy_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.191354] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec1fb0e-eae1-4652-9394-8b99a4dc7652 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.200104] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68792106-7220-483d-acd6-f3aca8a0c671 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.232943] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27d3818-b043-4490-af5e-4d1eeafdfa61 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.241698] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a88304-9526-427e-8261-1270252e2203 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.257144] env[62914]: DEBUG nova.compute.provider_tree [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.441670] env[62914]: DEBUG oslo_vmware.api [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352831, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228265} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.442016] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 954.442173] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 954.442360] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 954.442552] env[62914]: INFO nova.compute.manager [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Took 1.14 seconds to destroy the instance on the hypervisor. [ 954.442847] env[62914]: DEBUG oslo.service.loopingcall [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.443108] env[62914]: DEBUG nova.compute.manager [-] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 954.443208] env[62914]: DEBUG nova.network.neutron [-] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 954.511665] env[62914]: DEBUG oslo_concurrency.lockutils [None req-20616e04-89c0-43f8-925e-862e3ebfd8eb tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 25.756s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.622593] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352832, 'name': Destroy_Task, 'duration_secs': 0.394337} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.622923] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Destroyed the VM [ 954.623204] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Deleting Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 954.623482] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-002c31d1-8d72-48df-8dd6-c552f4ae2aeb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.632525] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 954.632525] env[62914]: value = "task-1352833" [ 954.632525] env[62914]: _type = "Task" [ 954.632525] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.642553] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352833, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.758170] env[62914]: DEBUG nova.compute.manager [req-5d0966b3-5c2f-4c6c-bbc7-6f8a843abdf3 req-a570b6ee-174c-4550-8473-4ee271e3b60e service nova] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Received event network-vif-deleted-0c5a4114-78d6-4319-9b49-f26a44441811 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.758325] env[62914]: INFO nova.compute.manager [req-5d0966b3-5c2f-4c6c-bbc7-6f8a843abdf3 req-a570b6ee-174c-4550-8473-4ee271e3b60e service nova] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Neutron deleted interface 0c5a4114-78d6-4319-9b49-f26a44441811; detaching it from the instance and deleting it from the info cache [ 954.758503] env[62914]: DEBUG nova.network.neutron [req-5d0966b3-5c2f-4c6c-bbc7-6f8a843abdf3 req-a570b6ee-174c-4550-8473-4ee271e3b60e service nova] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.760719] env[62914]: DEBUG nova.scheduler.client.report [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 955.044594] env[62914]: DEBUG nova.compute.manager [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 955.068923] env[62914]: DEBUG nova.virt.hardware [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 955.069232] env[62914]: DEBUG nova.virt.hardware [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 955.069433] env[62914]: DEBUG nova.virt.hardware [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 955.069677] env[62914]: DEBUG nova.virt.hardware [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 955.069884] env[62914]: DEBUG nova.virt.hardware [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 955.070186] env[62914]: DEBUG nova.virt.hardware [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 955.070452] env[62914]: DEBUG nova.virt.hardware [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 955.070647] env[62914]: DEBUG nova.virt.hardware [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 955.072299] env[62914]: DEBUG nova.virt.hardware [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 955.072547] env[62914]: DEBUG nova.virt.hardware [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 955.072786] env[62914]: DEBUG nova.virt.hardware [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 955.073706] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b42269-da3d-453e-a90e-e034a220965d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.082951] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc9a291-c4d1-4597-9871-8b73bb83a7e1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.143498] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352833, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.230254] env[62914]: DEBUG nova.network.neutron [-] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.262658] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d6a0c7c0-bc9d-458b-8c0f-64db3617ad6d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.264707] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.247s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.267177] env[62914]: DEBUG oslo_concurrency.lockutils [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.078s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.267446] env[62914]: DEBUG nova.objects.instance [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lazy-loading 'resources' on Instance uuid ab036e1f-a62a-490a-818c-2e1256c781f6 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.274752] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0beb251d-04f4-48a3-9222-616198074d3a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.309920] env[62914]: DEBUG nova.compute.manager [req-5d0966b3-5c2f-4c6c-bbc7-6f8a843abdf3 req-a570b6ee-174c-4550-8473-4ee271e3b60e service nova] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Detach interface failed, port_id=0c5a4114-78d6-4319-9b49-f26a44441811, reason: Instance 10c348bd-d85d-49b8-9b0f-216142f664d1 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 955.474213] env[62914]: DEBUG nova.compute.manager [req-d41c4388-3083-466d-8638-c0795199c28a req-3b71dcd4-26ee-42ab-9ccb-c9c64ef4288c service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Received event network-vif-plugged-e02166ec-c4c0-44a5-a9d6-5927b987648d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.474438] env[62914]: DEBUG oslo_concurrency.lockutils [req-d41c4388-3083-466d-8638-c0795199c28a req-3b71dcd4-26ee-42ab-9ccb-c9c64ef4288c service nova] Acquiring lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.474651] env[62914]: DEBUG oslo_concurrency.lockutils [req-d41c4388-3083-466d-8638-c0795199c28a req-3b71dcd4-26ee-42ab-9ccb-c9c64ef4288c service nova] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.474815] env[62914]: DEBUG oslo_concurrency.lockutils [req-d41c4388-3083-466d-8638-c0795199c28a req-3b71dcd4-26ee-42ab-9ccb-c9c64ef4288c service nova] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.475065] env[62914]: DEBUG nova.compute.manager [req-d41c4388-3083-466d-8638-c0795199c28a req-3b71dcd4-26ee-42ab-9ccb-c9c64ef4288c service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] No waiting events found dispatching network-vif-plugged-e02166ec-c4c0-44a5-a9d6-5927b987648d {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.475262] env[62914]: WARNING nova.compute.manager [req-d41c4388-3083-466d-8638-c0795199c28a req-3b71dcd4-26ee-42ab-9ccb-c9c64ef4288c service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Received unexpected event network-vif-plugged-e02166ec-c4c0-44a5-a9d6-5927b987648d for instance with vm_state building and task_state spawning. [ 955.556676] env[62914]: DEBUG nova.network.neutron [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Successfully updated port: e02166ec-c4c0-44a5-a9d6-5927b987648d {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.644323] env[62914]: DEBUG oslo_vmware.api [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352833, 'name': RemoveSnapshot_Task, 'duration_secs': 0.629534} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.644671] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Deleted Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 955.644960] env[62914]: INFO nova.compute.manager [None req-2f54787b-855d-4299-b29c-be96c5862723 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Took 15.54 seconds to snapshot the instance on the hypervisor. [ 955.733768] env[62914]: INFO nova.compute.manager [-] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Took 1.29 seconds to deallocate network for instance. [ 955.774256] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6d2c1cad-af75-42c3-a2b7-14ff8a7e2eb0 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "13954007-b5ee-4254-b99e-5fbb2f6e0add" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.437s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.775302] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ae834d04-897e-4198-a7f3-85fc110128c1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "13954007-b5ee-4254-b99e-5fbb2f6e0add" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 3.899s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.775521] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ae834d04-897e-4198-a7f3-85fc110128c1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "13954007-b5ee-4254-b99e-5fbb2f6e0add-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.775771] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ae834d04-897e-4198-a7f3-85fc110128c1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "13954007-b5ee-4254-b99e-5fbb2f6e0add-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.775884] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ae834d04-897e-4198-a7f3-85fc110128c1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "13954007-b5ee-4254-b99e-5fbb2f6e0add-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.777867] env[62914]: INFO nova.compute.manager [None req-ae834d04-897e-4198-a7f3-85fc110128c1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Terminating instance [ 955.779537] env[62914]: DEBUG nova.compute.manager [None req-ae834d04-897e-4198-a7f3-85fc110128c1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 955.779780] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ae834d04-897e-4198-a7f3-85fc110128c1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.780090] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f44e5d7-776f-49da-b476-2e1b4ce5899c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.792541] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bead3a5-ae50-46e7-8878-3cb2a8458aae {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.824589] env[62914]: WARNING nova.virt.vmwareapi.vmops [None req-ae834d04-897e-4198-a7f3-85fc110128c1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 13954007-b5ee-4254-b99e-5fbb2f6e0add could not be found. [ 955.824709] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-ae834d04-897e-4198-a7f3-85fc110128c1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 955.824898] env[62914]: INFO nova.compute.manager [None req-ae834d04-897e-4198-a7f3-85fc110128c1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Took 0.05 seconds to destroy the instance on the hypervisor. [ 955.825200] env[62914]: DEBUG oslo.service.loopingcall [None req-ae834d04-897e-4198-a7f3-85fc110128c1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.825435] env[62914]: DEBUG nova.compute.manager [-] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 955.825534] env[62914]: DEBUG nova.network.neutron [-] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 955.917582] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464d6801-3c14-4e79-a4aa-e00ba45febef {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.926369] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63230fe7-d08b-47af-a583-1e8ee196e635 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.961142] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c0f18ec-df18-453c-8c89-40a4b29224d5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.966659] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "a2a13ec9-30f3-4c04-b10b-f9a9c01534bd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.966923] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "a2a13ec9-30f3-4c04-b10b-f9a9c01534bd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.974363] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d8bfa9-dc88-42ab-9557-79844ed01e79 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.989867] env[62914]: DEBUG nova.compute.provider_tree [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.058822] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.058982] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquired lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.059156] env[62914]: DEBUG nova.network.neutron [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 956.240760] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.469468] env[62914]: DEBUG nova.compute.manager [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 956.492976] env[62914]: DEBUG nova.scheduler.client.report [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.536061] env[62914]: DEBUG nova.network.neutron [-] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.592013] env[62914]: DEBUG nova.network.neutron [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 956.735876] env[62914]: DEBUG nova.network.neutron [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Updating instance_info_cache with network_info: [{"id": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "address": "fa:16:3e:e8:d6:2b", "network": {"id": "74764d0b-64d3-4552-9517-ca2e87251c0b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1459372326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bb65abdffe44acdac0cdb8701e3e27b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape02166ec-c4", "ovs_interfaceid": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.993229] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.998095] env[62914]: DEBUG oslo_concurrency.lockutils [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.731s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.000074] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.760s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.000308] env[62914]: DEBUG nova.objects.instance [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lazy-loading 'resources' on Instance uuid 10c348bd-d85d-49b8-9b0f-216142f664d1 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.015588] env[62914]: INFO nova.scheduler.client.report [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Deleted allocations for instance ab036e1f-a62a-490a-818c-2e1256c781f6 [ 957.039028] env[62914]: INFO nova.compute.manager [-] [instance: 13954007-b5ee-4254-b99e-5fbb2f6e0add] Took 1.21 seconds to deallocate network for instance. [ 957.238584] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Releasing lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.238852] env[62914]: DEBUG nova.compute.manager [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Instance network_info: |[{"id": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "address": "fa:16:3e:e8:d6:2b", "network": {"id": "74764d0b-64d3-4552-9517-ca2e87251c0b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1459372326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bb65abdffe44acdac0cdb8701e3e27b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape02166ec-c4", "ovs_interfaceid": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 957.239325] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:d6:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '31e77685-b4dd-4810-80ef-24115ea9ea62', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e02166ec-c4c0-44a5-a9d6-5927b987648d', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 957.246804] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Creating folder: Project (7bb65abdffe44acdac0cdb8701e3e27b). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 957.247082] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6212e6cd-2212-4524-a725-6edd96bf8ef4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.258946] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Created folder: Project (7bb65abdffe44acdac0cdb8701e3e27b) in parent group-v288131. [ 957.259162] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Creating folder: Instances. Parent ref: group-v288251. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 957.259404] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7bb7aafb-98a3-4c3b-b59b-fb31a60ac26e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.270110] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Created folder: Instances in parent group-v288251. [ 957.270369] env[62914]: DEBUG oslo.service.loopingcall [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.270763] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 957.270763] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1144119e-001a-4af9-b68e-6233b4d4b7fd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.290789] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 957.290789] env[62914]: value = "task-1352836" [ 957.290789] env[62914]: _type = "Task" [ 957.290789] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.299989] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352836, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.503365] env[62914]: DEBUG nova.compute.manager [req-0f96d8ba-c908-4d92-ad87-9876152f2ada req-951d49ea-d6c9-4ec3-a0a2-499a84a5e526 service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Received event network-changed-e02166ec-c4c0-44a5-a9d6-5927b987648d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 957.503621] env[62914]: DEBUG nova.compute.manager [req-0f96d8ba-c908-4d92-ad87-9876152f2ada req-951d49ea-d6c9-4ec3-a0a2-499a84a5e526 service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Refreshing instance network info cache due to event network-changed-e02166ec-c4c0-44a5-a9d6-5927b987648d. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 957.503662] env[62914]: DEBUG oslo_concurrency.lockutils [req-0f96d8ba-c908-4d92-ad87-9876152f2ada req-951d49ea-d6c9-4ec3-a0a2-499a84a5e526 service nova] Acquiring lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.503890] env[62914]: DEBUG oslo_concurrency.lockutils [req-0f96d8ba-c908-4d92-ad87-9876152f2ada req-951d49ea-d6c9-4ec3-a0a2-499a84a5e526 service nova] Acquired lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.504020] env[62914]: DEBUG nova.network.neutron [req-0f96d8ba-c908-4d92-ad87-9876152f2ada req-951d49ea-d6c9-4ec3-a0a2-499a84a5e526 service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Refreshing network info cache for port e02166ec-c4c0-44a5-a9d6-5927b987648d {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 957.524953] env[62914]: DEBUG oslo_concurrency.lockutils [None req-693e7b6c-3ccb-45a1-a398-c896bd0059f6 tempest-ListServerFiltersTestJSON-1486379285 tempest-ListServerFiltersTestJSON-1486379285-project-member] Lock "ab036e1f-a62a-490a-818c-2e1256c781f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.243s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.623860] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de587bc0-8da5-4d26-8930-530987c0b344 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.634127] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef8cf3f-60be-4914-908d-3560ca4bf5bb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.665322] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3720bd84-3596-4238-89ee-6051190d7677 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.676192] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f59255-5f94-4d0b-9e59-8e65106598c6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.689385] env[62914]: DEBUG nova.compute.provider_tree [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.800744] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352836, 'name': CreateVM_Task, 'duration_secs': 0.386589} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.800920] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 957.801654] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.801854] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.802166] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 957.802428] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5dc8f7aa-7ae5-4631-88ba-8e97f23570d5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.808022] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 957.808022] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52883aab-3c95-0f17-45cc-f830c6be29fb" [ 957.808022] env[62914]: _type = "Task" [ 957.808022] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.817697] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52883aab-3c95-0f17-45cc-f830c6be29fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.062701] env[62914]: DEBUG oslo_concurrency.lockutils [None req-ae834d04-897e-4198-a7f3-85fc110128c1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "13954007-b5ee-4254-b99e-5fbb2f6e0add" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.287s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.192143] env[62914]: DEBUG nova.scheduler.client.report [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.213239] env[62914]: DEBUG nova.network.neutron [req-0f96d8ba-c908-4d92-ad87-9876152f2ada req-951d49ea-d6c9-4ec3-a0a2-499a84a5e526 service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Updated VIF entry in instance network info cache for port e02166ec-c4c0-44a5-a9d6-5927b987648d. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 958.213623] env[62914]: DEBUG nova.network.neutron [req-0f96d8ba-c908-4d92-ad87-9876152f2ada req-951d49ea-d6c9-4ec3-a0a2-499a84a5e526 service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Updating instance_info_cache with network_info: [{"id": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "address": "fa:16:3e:e8:d6:2b", "network": {"id": "74764d0b-64d3-4552-9517-ca2e87251c0b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1459372326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bb65abdffe44acdac0cdb8701e3e27b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape02166ec-c4", "ovs_interfaceid": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.319611] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52883aab-3c95-0f17-45cc-f830c6be29fb, 'name': SearchDatastore_Task, 'duration_secs': 0.013339} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.319929] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.320226] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 958.320506] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.320658] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.320838] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 958.321125] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6891478f-f0e4-43e9-9a95-25f952303a32 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.330266] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 958.330451] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 958.331168] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b49c853c-7a57-4636-b8ac-a2ad194462c9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.337248] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 958.337248] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c63617-9f35-b034-d99f-91e21032d88d" [ 958.337248] env[62914]: _type = "Task" [ 958.337248] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.345957] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c63617-9f35-b034-d99f-91e21032d88d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.696942] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.696s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.699834] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.707s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.701393] env[62914]: INFO nova.compute.claims [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.714224] env[62914]: INFO nova.scheduler.client.report [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleted allocations for instance 10c348bd-d85d-49b8-9b0f-216142f664d1 [ 958.716405] env[62914]: DEBUG oslo_concurrency.lockutils [req-0f96d8ba-c908-4d92-ad87-9876152f2ada req-951d49ea-d6c9-4ec3-a0a2-499a84a5e526 service nova] Releasing lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.848470] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c63617-9f35-b034-d99f-91e21032d88d, 'name': SearchDatastore_Task, 'duration_secs': 0.008973} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.849212] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0fde715f-fe86-4e53-85b9-6b9307ba14d0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.854339] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 958.854339] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52dd43b2-b72a-885c-0882-976abf5be7dd" [ 958.854339] env[62914]: _type = "Task" [ 958.854339] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.861464] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52dd43b2-b72a-885c-0882-976abf5be7dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.027566] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "5e8f452f-68e6-4fe6-9955-20359d90851b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.027807] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "5e8f452f-68e6-4fe6-9955-20359d90851b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.223305] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9f23cacc-ca22-46a4-82ad-65888cb06539 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "10c348bd-d85d-49b8-9b0f-216142f664d1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.928s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.365418] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52dd43b2-b72a-885c-0882-976abf5be7dd, 'name': SearchDatastore_Task, 'duration_secs': 0.009071} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.366020] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.366340] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 062cd4d7-3cf1-4f31-9631-e9d1e194ad53/062cd4d7-3cf1-4f31-9631-e9d1e194ad53.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 959.366644] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-64131e71-9e7d-4329-8a06-b82140198769 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.373634] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 959.373634] env[62914]: value = "task-1352837" [ 959.373634] env[62914]: _type = "Task" [ 959.373634] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.381897] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352837, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.530134] env[62914]: DEBUG nova.compute.manager [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 959.825768] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7eb185e-8466-45fd-b1d5-d13dc58592ac {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.836636] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aebcc7e-db23-43e2-8ba5-203f5d92a4d7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.867308] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4588aba-16bb-4832-bdd3-fec0ec68475c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.874715] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aacada3f-6c12-4799-9252-8bc2dfb93d44 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.890319] env[62914]: DEBUG nova.compute.provider_tree [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.894392] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352837, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483881} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.894824] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 062cd4d7-3cf1-4f31-9631-e9d1e194ad53/062cd4d7-3cf1-4f31-9631-e9d1e194ad53.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 959.895286] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 959.895286] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c93762b7-1f3c-4c56-a0e1-bd8c312eb0c9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.901695] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 959.901695] env[62914]: value = "task-1352838" [ 959.901695] env[62914]: _type = "Task" [ 959.901695] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.911690] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352838, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.953074] env[62914]: DEBUG oslo_concurrency.lockutils [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "4bec4c88-bd1e-4cdf-9c8a-eee56088479b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.953074] env[62914]: DEBUG oslo_concurrency.lockutils [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "4bec4c88-bd1e-4cdf-9c8a-eee56088479b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.953074] env[62914]: DEBUG oslo_concurrency.lockutils [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "4bec4c88-bd1e-4cdf-9c8a-eee56088479b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.953074] env[62914]: DEBUG oslo_concurrency.lockutils [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "4bec4c88-bd1e-4cdf-9c8a-eee56088479b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.953351] env[62914]: DEBUG oslo_concurrency.lockutils [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "4bec4c88-bd1e-4cdf-9c8a-eee56088479b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.955860] env[62914]: INFO nova.compute.manager [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Terminating instance [ 959.957966] env[62914]: DEBUG nova.compute.manager [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 959.958458] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 959.959136] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ff99da-0c2c-4e59-9f7d-1bfe1ce897cc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.970350] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 959.970616] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4e49389a-f084-4ba8-b722-602a18eb46c6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.978447] env[62914]: DEBUG oslo_vmware.api [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 959.978447] env[62914]: value = "task-1352839" [ 959.978447] env[62914]: _type = "Task" [ 959.978447] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.988407] env[62914]: DEBUG oslo_vmware.api [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352839, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.048322] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.396840] env[62914]: DEBUG nova.scheduler.client.report [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.412248] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352838, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060068} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.413309] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 960.414093] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30737d35-00a9-4186-b009-ba5374e43771 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.436576] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 062cd4d7-3cf1-4f31-9631-e9d1e194ad53/062cd4d7-3cf1-4f31-9631-e9d1e194ad53.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.437077] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53c88db8-ef68-47a3-a22c-0e81784da924 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.457194] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 960.457194] env[62914]: value = "task-1352840" [ 960.457194] env[62914]: _type = "Task" [ 960.457194] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.465452] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352840, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.487459] env[62914]: DEBUG oslo_vmware.api [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352839, 'name': PowerOffVM_Task, 'duration_secs': 0.19278} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.487682] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 960.487841] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 960.488127] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-952a0e81-071d-4269-9b8d-07445848f1c7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.580370] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 960.580618] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 960.580807] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleting the datastore file [datastore1] 4bec4c88-bd1e-4cdf-9c8a-eee56088479b {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.581118] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f9bc9b8f-3ab0-48a1-a262-030d22edcaba {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.587606] env[62914]: DEBUG oslo_vmware.api [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 960.587606] env[62914]: value = "task-1352842" [ 960.587606] env[62914]: _type = "Task" [ 960.587606] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.597997] env[62914]: DEBUG oslo_vmware.api [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352842, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.902219] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.202s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.902731] env[62914]: DEBUG nova.compute.manager [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 960.905694] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.857s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.907057] env[62914]: INFO nova.compute.claims [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.968854] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352840, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.098251] env[62914]: DEBUG oslo_vmware.api [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352842, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135082} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.098520] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.098707] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 961.098891] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 961.099084] env[62914]: INFO nova.compute.manager [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 961.099330] env[62914]: DEBUG oslo.service.loopingcall [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.099528] env[62914]: DEBUG nova.compute.manager [-] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 961.099623] env[62914]: DEBUG nova.network.neutron [-] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 961.413666] env[62914]: DEBUG nova.compute.utils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 961.494194] env[62914]: DEBUG nova.compute.manager [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 961.494194] env[62914]: DEBUG nova.network.neutron [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 961.494194] env[62914]: DEBUG nova.compute.manager [req-4a621f6a-1dfd-485e-85e6-898f22205c15 req-180f62f4-ad81-4440-92c9-786cdce0fb2e service nova] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Received event network-vif-deleted-79e0436a-5077-4b43-8967-969e5a2e3799 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.494194] env[62914]: INFO nova.compute.manager [req-4a621f6a-1dfd-485e-85e6-898f22205c15 req-180f62f4-ad81-4440-92c9-786cdce0fb2e service nova] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Neutron deleted interface 79e0436a-5077-4b43-8967-969e5a2e3799; detaching it from the instance and deleting it from the info cache [ 961.494194] env[62914]: DEBUG nova.network.neutron [req-4a621f6a-1dfd-485e-85e6-898f22205c15 req-180f62f4-ad81-4440-92c9-786cdce0fb2e service nova] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.494194] env[62914]: DEBUG nova.policy [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41c19b40819c49d8a9dacdb74ad5899b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f5fa79f0260e461c952301f904101f79', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 961.494194] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352840, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.751046] env[62914]: DEBUG nova.network.neutron [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Successfully created port: c37fe151-4337-4f57-855a-5b250e70ce13 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 961.892741] env[62914]: DEBUG nova.network.neutron [-] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.925037] env[62914]: DEBUG nova.compute.manager [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 961.929012] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d956ea1f-3d83-4e48-8d61-645ba0220ef2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.938548] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3546aee8-1b80-4e5a-abd6-feac2c96ca58 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.974423] env[62914]: DEBUG nova.compute.manager [req-4a621f6a-1dfd-485e-85e6-898f22205c15 req-180f62f4-ad81-4440-92c9-786cdce0fb2e service nova] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Detach interface failed, port_id=79e0436a-5077-4b43-8967-969e5a2e3799, reason: Instance 4bec4c88-bd1e-4cdf-9c8a-eee56088479b could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 961.978904] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352840, 'name': ReconfigVM_Task, 'duration_secs': 1.033455} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.979099] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 062cd4d7-3cf1-4f31-9631-e9d1e194ad53/062cd4d7-3cf1-4f31-9631-e9d1e194ad53.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.979724] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cf2ab476-1f35-45b6-a74f-a24691f136f2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.987559] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 961.987559] env[62914]: value = "task-1352843" [ 961.987559] env[62914]: _type = "Task" [ 961.987559] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.999700] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352843, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.061884] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4cd4289-cc74-4a59-b8e9-abf7c2e70287 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.070141] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad333f3-0886-4771-92e5-330bcd45deec {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.101890] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-021949f5-33d2-4c91-88b7-1aea208e9109 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.109437] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f99718c-f0e3-442f-936e-d392770140cc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.123649] env[62914]: DEBUG nova.compute.provider_tree [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.395407] env[62914]: INFO nova.compute.manager [-] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Took 1.30 seconds to deallocate network for instance. [ 962.500216] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352843, 'name': Rename_Task, 'duration_secs': 0.142407} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.500572] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 962.500681] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e539aa5-ff12-4fc0-93ee-7f943027a8b6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.507537] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 962.507537] env[62914]: value = "task-1352844" [ 962.507537] env[62914]: _type = "Task" [ 962.507537] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.515331] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352844, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.626776] env[62914]: DEBUG nova.scheduler.client.report [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.901623] env[62914]: DEBUG oslo_concurrency.lockutils [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.935046] env[62914]: DEBUG nova.compute.manager [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 962.963930] env[62914]: DEBUG nova.virt.hardware [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='92e09b1207672e290b779373f340fbd9',container_format='bare',created_at=2024-10-24T12:15:56Z,direct_url=,disk_format='vmdk',id=67e90822-61d3-4074-a5de-33c47f66782e,min_disk=1,min_ram=0,name='tempest-test-snap-522476699',owner='f5fa79f0260e461c952301f904101f79',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-24T12:16:11Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.963930] env[62914]: DEBUG nova.virt.hardware [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.964155] env[62914]: DEBUG nova.virt.hardware [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.964240] env[62914]: DEBUG nova.virt.hardware [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.964426] env[62914]: DEBUG nova.virt.hardware [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.964576] env[62914]: DEBUG nova.virt.hardware [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.964783] env[62914]: DEBUG nova.virt.hardware [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.964945] env[62914]: DEBUG nova.virt.hardware [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.965138] env[62914]: DEBUG nova.virt.hardware [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.965302] env[62914]: DEBUG nova.virt.hardware [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.965475] env[62914]: DEBUG nova.virt.hardware [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.966336] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dfe0b38-4923-4850-94ab-98df61f8113f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.974752] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0767a52f-c6b1-47ba-8c15-4c9c58f43ca4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.017752] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352844, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.132020] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.226s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.132572] env[62914]: DEBUG nova.compute.manager [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 963.135332] env[62914]: DEBUG oslo_concurrency.lockutils [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.234s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.135614] env[62914]: DEBUG nova.objects.instance [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lazy-loading 'resources' on Instance uuid 4bec4c88-bd1e-4cdf-9c8a-eee56088479b {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.216095] env[62914]: DEBUG nova.network.neutron [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Successfully updated port: c37fe151-4337-4f57-855a-5b250e70ce13 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 963.518041] env[62914]: DEBUG oslo_vmware.api [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352844, 'name': PowerOnVM_Task, 'duration_secs': 0.854342} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.518919] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 963.519152] env[62914]: INFO nova.compute.manager [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Took 8.47 seconds to spawn the instance on the hypervisor. [ 963.519382] env[62914]: DEBUG nova.compute.manager [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 963.520999] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964b7f50-2d0a-4696-a097-7576f461d0ef {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.524744] env[62914]: DEBUG nova.compute.manager [req-8150c3f5-9ce2-448b-9e37-925493d4d04d req-d8ce1706-ed01-40ac-a64b-02aa4476c050 service nova] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Received event network-vif-plugged-c37fe151-4337-4f57-855a-5b250e70ce13 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.524951] env[62914]: DEBUG oslo_concurrency.lockutils [req-8150c3f5-9ce2-448b-9e37-925493d4d04d req-d8ce1706-ed01-40ac-a64b-02aa4476c050 service nova] Acquiring lock "a2a13ec9-30f3-4c04-b10b-f9a9c01534bd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.525178] env[62914]: DEBUG oslo_concurrency.lockutils [req-8150c3f5-9ce2-448b-9e37-925493d4d04d req-d8ce1706-ed01-40ac-a64b-02aa4476c050 service nova] Lock "a2a13ec9-30f3-4c04-b10b-f9a9c01534bd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.525358] env[62914]: DEBUG oslo_concurrency.lockutils [req-8150c3f5-9ce2-448b-9e37-925493d4d04d req-d8ce1706-ed01-40ac-a64b-02aa4476c050 service nova] Lock "a2a13ec9-30f3-4c04-b10b-f9a9c01534bd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.525532] env[62914]: DEBUG nova.compute.manager [req-8150c3f5-9ce2-448b-9e37-925493d4d04d req-d8ce1706-ed01-40ac-a64b-02aa4476c050 service nova] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] No waiting events found dispatching network-vif-plugged-c37fe151-4337-4f57-855a-5b250e70ce13 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 963.525797] env[62914]: WARNING nova.compute.manager [req-8150c3f5-9ce2-448b-9e37-925493d4d04d req-d8ce1706-ed01-40ac-a64b-02aa4476c050 service nova] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Received unexpected event network-vif-plugged-c37fe151-4337-4f57-855a-5b250e70ce13 for instance with vm_state building and task_state spawning. [ 963.526090] env[62914]: DEBUG nova.compute.manager [req-8150c3f5-9ce2-448b-9e37-925493d4d04d req-d8ce1706-ed01-40ac-a64b-02aa4476c050 service nova] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Received event network-changed-c37fe151-4337-4f57-855a-5b250e70ce13 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.526370] env[62914]: DEBUG nova.compute.manager [req-8150c3f5-9ce2-448b-9e37-925493d4d04d req-d8ce1706-ed01-40ac-a64b-02aa4476c050 service nova] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Refreshing instance network info cache due to event network-changed-c37fe151-4337-4f57-855a-5b250e70ce13. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 963.526677] env[62914]: DEBUG oslo_concurrency.lockutils [req-8150c3f5-9ce2-448b-9e37-925493d4d04d req-d8ce1706-ed01-40ac-a64b-02aa4476c050 service nova] Acquiring lock "refresh_cache-a2a13ec9-30f3-4c04-b10b-f9a9c01534bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.526908] env[62914]: DEBUG oslo_concurrency.lockutils [req-8150c3f5-9ce2-448b-9e37-925493d4d04d req-d8ce1706-ed01-40ac-a64b-02aa4476c050 service nova] Acquired lock "refresh_cache-a2a13ec9-30f3-4c04-b10b-f9a9c01534bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.527113] env[62914]: DEBUG nova.network.neutron [req-8150c3f5-9ce2-448b-9e37-925493d4d04d req-d8ce1706-ed01-40ac-a64b-02aa4476c050 service nova] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Refreshing network info cache for port c37fe151-4337-4f57-855a-5b250e70ce13 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 963.639024] env[62914]: DEBUG nova.compute.utils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 963.643694] env[62914]: DEBUG nova.compute.manager [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 963.643694] env[62914]: DEBUG nova.network.neutron [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 963.683011] env[62914]: DEBUG nova.policy [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2c21c59e39b94deea6c3701fe971198a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1dcda44f0d6a455bb17966e7c5061190', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 963.719149] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "refresh_cache-a2a13ec9-30f3-4c04-b10b-f9a9c01534bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.742415] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72e16a9-f6a0-47c7-8b54-272789ed0764 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.750080] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c503eea-1c7c-4bb7-9638-8076b1ea74ca {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.780469] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e13b820c-c0fc-4175-98d3-5ea05f1a9f7c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.788391] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed02e7ff-400e-4711-86de-6171ab28e000 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.801774] env[62914]: DEBUG nova.compute.provider_tree [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.895127] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f85f0b23-09ac-4358-8853-b83c4917a030 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "interface-1e527707-758c-4006-af83-9c739b9645ed-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.895127] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f85f0b23-09ac-4358-8853-b83c4917a030 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-1e527707-758c-4006-af83-9c739b9645ed-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.895127] env[62914]: DEBUG nova.objects.instance [None req-f85f0b23-09ac-4358-8853-b83c4917a030 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'flavor' on Instance uuid 1e527707-758c-4006-af83-9c739b9645ed {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.929309] env[62914]: DEBUG nova.network.neutron [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Successfully created port: 90f668f6-799d-4a5a-a911-23dad79f3c4d {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 964.044921] env[62914]: INFO nova.compute.manager [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Took 13.27 seconds to build instance. [ 964.072484] env[62914]: DEBUG nova.network.neutron [req-8150c3f5-9ce2-448b-9e37-925493d4d04d req-d8ce1706-ed01-40ac-a64b-02aa4476c050 service nova] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 964.144456] env[62914]: DEBUG nova.compute.manager [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 964.148136] env[62914]: DEBUG nova.network.neutron [req-8150c3f5-9ce2-448b-9e37-925493d4d04d req-d8ce1706-ed01-40ac-a64b-02aa4476c050 service nova] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.304589] env[62914]: DEBUG nova.scheduler.client.report [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.399671] env[62914]: DEBUG nova.objects.instance [None req-f85f0b23-09ac-4358-8853-b83c4917a030 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'pci_requests' on Instance uuid 1e527707-758c-4006-af83-9c739b9645ed {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.546945] env[62914]: DEBUG oslo_concurrency.lockutils [None req-286ad467-5e4d-4baf-9a3a-067b5088d380 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.789s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.653545] env[62914]: DEBUG oslo_concurrency.lockutils [req-8150c3f5-9ce2-448b-9e37-925493d4d04d req-d8ce1706-ed01-40ac-a64b-02aa4476c050 service nova] Releasing lock "refresh_cache-a2a13ec9-30f3-4c04-b10b-f9a9c01534bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.653935] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "refresh_cache-a2a13ec9-30f3-4c04-b10b-f9a9c01534bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.654138] env[62914]: DEBUG nova.network.neutron [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 964.810044] env[62914]: DEBUG oslo_concurrency.lockutils [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.674s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.834757] env[62914]: INFO nova.scheduler.client.report [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleted allocations for instance 4bec4c88-bd1e-4cdf-9c8a-eee56088479b [ 964.902475] env[62914]: DEBUG nova.objects.base [None req-f85f0b23-09ac-4358-8853-b83c4917a030 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Object Instance<1e527707-758c-4006-af83-9c739b9645ed> lazy-loaded attributes: flavor,pci_requests {{(pid=62914) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 964.902718] env[62914]: DEBUG nova.network.neutron [None req-f85f0b23-09ac-4358-8853-b83c4917a030 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 965.042853] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f85f0b23-09ac-4358-8853-b83c4917a030 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-1e527707-758c-4006-af83-9c739b9645ed-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.148s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.160027] env[62914]: DEBUG nova.compute.manager [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 965.189698] env[62914]: DEBUG nova.virt.hardware [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 965.189698] env[62914]: DEBUG nova.virt.hardware [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 965.189698] env[62914]: DEBUG nova.virt.hardware [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.189946] env[62914]: DEBUG nova.virt.hardware [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 965.189946] env[62914]: DEBUG nova.virt.hardware [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.190067] env[62914]: DEBUG nova.virt.hardware [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 965.190281] env[62914]: DEBUG nova.virt.hardware [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 965.190506] env[62914]: DEBUG nova.virt.hardware [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 965.190685] env[62914]: DEBUG nova.virt.hardware [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 965.190852] env[62914]: DEBUG nova.virt.hardware [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 965.191036] env[62914]: DEBUG nova.virt.hardware [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.191955] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d3beae-7e85-4cae-9260-0a99bf430ae4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.200232] env[62914]: DEBUG nova.network.neutron [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 965.203108] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5bdd5d2-c807-49a0-9c2f-9003f4e95911 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.343012] env[62914]: DEBUG oslo_concurrency.lockutils [None req-031e1a01-eb48-4844-a40a-3184b533ed79 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "4bec4c88-bd1e-4cdf-9c8a-eee56088479b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.390s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.410757] env[62914]: DEBUG nova.network.neutron [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Successfully updated port: 90f668f6-799d-4a5a-a911-23dad79f3c4d {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 965.508163] env[62914]: DEBUG nova.network.neutron [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Updating instance_info_cache with network_info: [{"id": "c37fe151-4337-4f57-855a-5b250e70ce13", "address": "fa:16:3e:86:c6:c1", "network": {"id": "c9c30d64-cafe-48a5-b868-b20d90d8920d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-458611021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5fa79f0260e461c952301f904101f79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc37fe151-43", "ovs_interfaceid": "c37fe151-4337-4f57-855a-5b250e70ce13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.644745] env[62914]: DEBUG nova.compute.manager [req-7d32f8f7-66cb-40c4-97d1-bd374c6d0267 req-06570357-30e4-45f2-8bf2-48551eeb4296 service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Received event network-changed-e02166ec-c4c0-44a5-a9d6-5927b987648d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.645044] env[62914]: DEBUG nova.compute.manager [req-7d32f8f7-66cb-40c4-97d1-bd374c6d0267 req-06570357-30e4-45f2-8bf2-48551eeb4296 service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Refreshing instance network info cache due to event network-changed-e02166ec-c4c0-44a5-a9d6-5927b987648d. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 965.645207] env[62914]: DEBUG oslo_concurrency.lockutils [req-7d32f8f7-66cb-40c4-97d1-bd374c6d0267 req-06570357-30e4-45f2-8bf2-48551eeb4296 service nova] Acquiring lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.645352] env[62914]: DEBUG oslo_concurrency.lockutils [req-7d32f8f7-66cb-40c4-97d1-bd374c6d0267 req-06570357-30e4-45f2-8bf2-48551eeb4296 service nova] Acquired lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.646072] env[62914]: DEBUG nova.network.neutron [req-7d32f8f7-66cb-40c4-97d1-bd374c6d0267 req-06570357-30e4-45f2-8bf2-48551eeb4296 service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Refreshing network info cache for port e02166ec-c4c0-44a5-a9d6-5927b987648d {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.665273] env[62914]: DEBUG nova.compute.manager [req-9f723bd4-5365-4350-9d52-ff0fa4b86b67 req-1238c457-c64c-4afe-abc0-2a5bb7aba655 service nova] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Received event network-vif-plugged-90f668f6-799d-4a5a-a911-23dad79f3c4d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.665513] env[62914]: DEBUG oslo_concurrency.lockutils [req-9f723bd4-5365-4350-9d52-ff0fa4b86b67 req-1238c457-c64c-4afe-abc0-2a5bb7aba655 service nova] Acquiring lock "5e8f452f-68e6-4fe6-9955-20359d90851b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.665738] env[62914]: DEBUG oslo_concurrency.lockutils [req-9f723bd4-5365-4350-9d52-ff0fa4b86b67 req-1238c457-c64c-4afe-abc0-2a5bb7aba655 service nova] Lock "5e8f452f-68e6-4fe6-9955-20359d90851b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.665926] env[62914]: DEBUG oslo_concurrency.lockutils [req-9f723bd4-5365-4350-9d52-ff0fa4b86b67 req-1238c457-c64c-4afe-abc0-2a5bb7aba655 service nova] Lock "5e8f452f-68e6-4fe6-9955-20359d90851b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.666144] env[62914]: DEBUG nova.compute.manager [req-9f723bd4-5365-4350-9d52-ff0fa4b86b67 req-1238c457-c64c-4afe-abc0-2a5bb7aba655 service nova] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] No waiting events found dispatching network-vif-plugged-90f668f6-799d-4a5a-a911-23dad79f3c4d {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 965.666313] env[62914]: WARNING nova.compute.manager [req-9f723bd4-5365-4350-9d52-ff0fa4b86b67 req-1238c457-c64c-4afe-abc0-2a5bb7aba655 service nova] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Received unexpected event network-vif-plugged-90f668f6-799d-4a5a-a911-23dad79f3c4d for instance with vm_state building and task_state spawning. [ 965.666479] env[62914]: DEBUG nova.compute.manager [req-9f723bd4-5365-4350-9d52-ff0fa4b86b67 req-1238c457-c64c-4afe-abc0-2a5bb7aba655 service nova] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Received event network-changed-90f668f6-799d-4a5a-a911-23dad79f3c4d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.666668] env[62914]: DEBUG nova.compute.manager [req-9f723bd4-5365-4350-9d52-ff0fa4b86b67 req-1238c457-c64c-4afe-abc0-2a5bb7aba655 service nova] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Refreshing instance network info cache due to event network-changed-90f668f6-799d-4a5a-a911-23dad79f3c4d. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 965.666864] env[62914]: DEBUG oslo_concurrency.lockutils [req-9f723bd4-5365-4350-9d52-ff0fa4b86b67 req-1238c457-c64c-4afe-abc0-2a5bb7aba655 service nova] Acquiring lock "refresh_cache-5e8f452f-68e6-4fe6-9955-20359d90851b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.667013] env[62914]: DEBUG oslo_concurrency.lockutils [req-9f723bd4-5365-4350-9d52-ff0fa4b86b67 req-1238c457-c64c-4afe-abc0-2a5bb7aba655 service nova] Acquired lock "refresh_cache-5e8f452f-68e6-4fe6-9955-20359d90851b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.667312] env[62914]: DEBUG nova.network.neutron [req-9f723bd4-5365-4350-9d52-ff0fa4b86b67 req-1238c457-c64c-4afe-abc0-2a5bb7aba655 service nova] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Refreshing network info cache for port 90f668f6-799d-4a5a-a911-23dad79f3c4d {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.913895] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "refresh_cache-5e8f452f-68e6-4fe6-9955-20359d90851b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.010944] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "refresh_cache-a2a13ec9-30f3-4c04-b10b-f9a9c01534bd" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.011593] env[62914]: DEBUG nova.compute.manager [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Instance network_info: |[{"id": "c37fe151-4337-4f57-855a-5b250e70ce13", "address": "fa:16:3e:86:c6:c1", "network": {"id": "c9c30d64-cafe-48a5-b868-b20d90d8920d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-458611021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5fa79f0260e461c952301f904101f79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc37fe151-43", "ovs_interfaceid": "c37fe151-4337-4f57-855a-5b250e70ce13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 966.011988] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:c6:c1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c37fe151-4337-4f57-855a-5b250e70ce13', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 966.019588] env[62914]: DEBUG oslo.service.loopingcall [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.019859] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 966.020141] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8cd5e6a5-2b5f-4b56-8202-1506e7ddfd69 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.043299] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 966.043299] env[62914]: value = "task-1352845" [ 966.043299] env[62914]: _type = "Task" [ 966.043299] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.052879] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352845, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.072888] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 966.073206] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 966.239181] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.239423] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.421288] env[62914]: DEBUG nova.network.neutron [req-7d32f8f7-66cb-40c4-97d1-bd374c6d0267 req-06570357-30e4-45f2-8bf2-48551eeb4296 service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Updated VIF entry in instance network info cache for port e02166ec-c4c0-44a5-a9d6-5927b987648d. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 966.421288] env[62914]: DEBUG nova.network.neutron [req-7d32f8f7-66cb-40c4-97d1-bd374c6d0267 req-06570357-30e4-45f2-8bf2-48551eeb4296 service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Updating instance_info_cache with network_info: [{"id": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "address": "fa:16:3e:e8:d6:2b", "network": {"id": "74764d0b-64d3-4552-9517-ca2e87251c0b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1459372326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bb65abdffe44acdac0cdb8701e3e27b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape02166ec-c4", "ovs_interfaceid": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.423364] env[62914]: DEBUG nova.network.neutron [req-9f723bd4-5365-4350-9d52-ff0fa4b86b67 req-1238c457-c64c-4afe-abc0-2a5bb7aba655 service nova] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 966.503257] env[62914]: DEBUG nova.network.neutron [req-9f723bd4-5365-4350-9d52-ff0fa4b86b67 req-1238c457-c64c-4afe-abc0-2a5bb7aba655 service nova] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.554142] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352845, 'name': CreateVM_Task, 'duration_secs': 0.361232} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.554366] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 966.555034] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/67e90822-61d3-4074-a5de-33c47f66782e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.555219] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "[datastore2] devstack-image-cache_base/67e90822-61d3-4074-a5de-33c47f66782e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.555613] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/67e90822-61d3-4074-a5de-33c47f66782e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.555873] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f558df52-1440-491a-ba48-2e066525de8d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.560473] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 966.560473] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f4d1f1-3f5b-0aad-c283-d795df990336" [ 966.560473] env[62914]: _type = "Task" [ 966.560473] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.568517] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f4d1f1-3f5b-0aad-c283-d795df990336, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.578775] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 966.578927] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Starting heal instance info cache {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 966.579059] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Rebuilding the list of instances to heal {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 966.741952] env[62914]: DEBUG nova.compute.manager [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 966.926705] env[62914]: DEBUG oslo_concurrency.lockutils [req-7d32f8f7-66cb-40c4-97d1-bd374c6d0267 req-06570357-30e4-45f2-8bf2-48551eeb4296 service nova] Releasing lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.933542] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "interface-1e527707-758c-4006-af83-9c739b9645ed-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.933817] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-1e527707-758c-4006-af83-9c739b9645ed-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.934160] env[62914]: DEBUG nova.objects.instance [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'flavor' on Instance uuid 1e527707-758c-4006-af83-9c739b9645ed {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.006294] env[62914]: DEBUG oslo_concurrency.lockutils [req-9f723bd4-5365-4350-9d52-ff0fa4b86b67 req-1238c457-c64c-4afe-abc0-2a5bb7aba655 service nova] Releasing lock "refresh_cache-5e8f452f-68e6-4fe6-9955-20359d90851b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.006677] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "refresh_cache-5e8f452f-68e6-4fe6-9955-20359d90851b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.006841] env[62914]: DEBUG nova.network.neutron [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 967.071257] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "[datastore2] devstack-image-cache_base/67e90822-61d3-4074-a5de-33c47f66782e" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.071532] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Processing image 67e90822-61d3-4074-a5de-33c47f66782e {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 967.071776] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/67e90822-61d3-4074-a5de-33c47f66782e/67e90822-61d3-4074-a5de-33c47f66782e.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.071930] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "[datastore2] devstack-image-cache_base/67e90822-61d3-4074-a5de-33c47f66782e/67e90822-61d3-4074-a5de-33c47f66782e.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.072157] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 967.072419] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d4e02eba-68e2-49bd-94ed-e370133437c6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.081111] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 967.081299] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 967.083560] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 967.083697] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Skipping network cache update for instance because it is Building. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 967.085900] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b6b83f6-a1ff-40aa-98e5-231a40b25843 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.091854] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 967.091854] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52e165b6-6cd4-b703-b5d2-88ed1991ebab" [ 967.091854] env[62914]: _type = "Task" [ 967.091854] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.099322] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52e165b6-6cd4-b703-b5d2-88ed1991ebab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.113127] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.113275] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquired lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.113417] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Forcefully refreshing network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 967.113565] env[62914]: DEBUG nova.objects.instance [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lazy-loading 'info_cache' on Instance uuid c7cf71d7-b823-471f-949f-9ef2b539d972 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.263845] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.264127] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.265650] env[62914]: INFO nova.compute.claims [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.503915] env[62914]: DEBUG nova.objects.instance [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'pci_requests' on Instance uuid 1e527707-758c-4006-af83-9c739b9645ed {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.533888] env[62914]: DEBUG nova.network.neutron [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 967.601653] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Preparing fetch location {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 967.601918] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Fetch image to [datastore2] OSTACK_IMG_b69cb349-e00e-428c-8078-55c1c7577ae8/OSTACK_IMG_b69cb349-e00e-428c-8078-55c1c7577ae8.vmdk {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 967.602146] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Downloading stream optimized image 67e90822-61d3-4074-a5de-33c47f66782e to [datastore2] OSTACK_IMG_b69cb349-e00e-428c-8078-55c1c7577ae8/OSTACK_IMG_b69cb349-e00e-428c-8078-55c1c7577ae8.vmdk on the data store datastore2 as vApp {{(pid=62914) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 967.602336] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Downloading image file data 67e90822-61d3-4074-a5de-33c47f66782e to the ESX as VM named 'OSTACK_IMG_b69cb349-e00e-428c-8078-55c1c7577ae8' {{(pid=62914) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 967.656035] env[62914]: DEBUG nova.network.neutron [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Updating instance_info_cache with network_info: [{"id": "90f668f6-799d-4a5a-a911-23dad79f3c4d", "address": "fa:16:3e:b2:88:36", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90f668f6-79", "ovs_interfaceid": "90f668f6-799d-4a5a-a911-23dad79f3c4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.673545] env[62914]: DEBUG oslo_vmware.rw_handles [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 967.673545] env[62914]: value = "resgroup-9" [ 967.673545] env[62914]: _type = "ResourcePool" [ 967.673545] env[62914]: }. {{(pid=62914) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 967.674061] env[62914]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-6ed1e022-4d8c-4789-a700-252440742e27 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.695665] env[62914]: DEBUG oslo_vmware.rw_handles [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lease: (returnval){ [ 967.695665] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5262ec52-1f4b-1443-8f4b-cf5a5f26d3cf" [ 967.695665] env[62914]: _type = "HttpNfcLease" [ 967.695665] env[62914]: } obtained for vApp import into resource pool (val){ [ 967.695665] env[62914]: value = "resgroup-9" [ 967.695665] env[62914]: _type = "ResourcePool" [ 967.695665] env[62914]: }. {{(pid=62914) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 967.696029] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the lease: (returnval){ [ 967.696029] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5262ec52-1f4b-1443-8f4b-cf5a5f26d3cf" [ 967.696029] env[62914]: _type = "HttpNfcLease" [ 967.696029] env[62914]: } to be ready. {{(pid=62914) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 967.702080] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 967.702080] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5262ec52-1f4b-1443-8f4b-cf5a5f26d3cf" [ 967.702080] env[62914]: _type = "HttpNfcLease" [ 967.702080] env[62914]: } is initializing. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 968.006273] env[62914]: DEBUG nova.objects.base [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Object Instance<1e527707-758c-4006-af83-9c739b9645ed> lazy-loaded attributes: flavor,pci_requests {{(pid=62914) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 968.006662] env[62914]: DEBUG nova.network.neutron [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 968.047024] env[62914]: DEBUG nova.policy [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80089586cfda4cd2bdfc24d9f556239c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09d51090ac5d41fc966c7e132f288824', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 968.157450] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "refresh_cache-5e8f452f-68e6-4fe6-9955-20359d90851b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.157776] env[62914]: DEBUG nova.compute.manager [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Instance network_info: |[{"id": "90f668f6-799d-4a5a-a911-23dad79f3c4d", "address": "fa:16:3e:b2:88:36", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90f668f6-79", "ovs_interfaceid": "90f668f6-799d-4a5a-a911-23dad79f3c4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 968.158226] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b2:88:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a8c8175-1197-4f12-baac-ef6aba95f585', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '90f668f6-799d-4a5a-a911-23dad79f3c4d', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 968.167275] env[62914]: DEBUG oslo.service.loopingcall [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.169022] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 968.169022] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba820775-f302-480f-8225-a4bdf1e7fa30 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.191495] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 968.191495] env[62914]: value = "task-1352847" [ 968.191495] env[62914]: _type = "Task" [ 968.191495] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.202302] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352847, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.205782] env[62914]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 968.205782] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5262ec52-1f4b-1443-8f4b-cf5a5f26d3cf" [ 968.205782] env[62914]: _type = "HttpNfcLease" [ 968.205782] env[62914]: } is ready. {{(pid=62914) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 968.206121] env[62914]: DEBUG oslo_vmware.rw_handles [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 968.206121] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5262ec52-1f4b-1443-8f4b-cf5a5f26d3cf" [ 968.206121] env[62914]: _type = "HttpNfcLease" [ 968.206121] env[62914]: }. {{(pid=62914) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 968.206925] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d83b0a-084d-4a6e-98a0-c5aa94688817 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.214186] env[62914]: DEBUG oslo_vmware.rw_handles [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5221e259-f9ca-f567-e6f0-5f63c7cd4d9e/disk-0.vmdk from lease info. {{(pid=62914) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 968.214394] env[62914]: DEBUG oslo_vmware.rw_handles [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5221e259-f9ca-f567-e6f0-5f63c7cd4d9e/disk-0.vmdk. {{(pid=62914) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 968.284281] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-916c64e7-449c-4bd5-b2f2-8f0d3b51d528 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.337243] env[62914]: DEBUG nova.network.neutron [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Successfully created port: 3e7456bf-2578-404e-98da-f253c33ac065 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 968.406013] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83411647-028f-496d-9905-b1432d8ff838 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.415196] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459c87ed-a889-4cb6-b812-282069f6ebc4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.452072] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-241fe447-e45c-44d8-aa6d-e2ae9715a8b9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.462614] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40bbc14a-faa6-4e9b-8864-fbcb56f6e303 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.480654] env[62914]: DEBUG nova.compute.provider_tree [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.708111] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352847, 'name': CreateVM_Task, 'duration_secs': 0.357084} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.708342] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 968.709184] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.709437] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.709852] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 968.712143] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45a275c3-75b0-4857-8b87-14cd1f660a9a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.717565] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 968.717565] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ab5d21-2666-948d-79b7-89181a69c92b" [ 968.717565] env[62914]: _type = "Task" [ 968.717565] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.731150] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ab5d21-2666-948d-79b7-89181a69c92b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.853157] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updating instance_info_cache with network_info: [{"id": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "address": "fa:16:3e:91:13:35", "network": {"id": "c4298e11-ab16-4b37-be29-48fc406e74b0", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-695272838-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.236", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "15315f8e77d749a4b2b0211c3b22b546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe38bb7e-8bcb-419d-868f-0dc105c69651", "external-id": "nsx-vlan-transportzone-432", "segmentation_id": 432, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57172af7-5e", "ovs_interfaceid": "57172af7-5ed6-46d9-9d7b-3ef100c530f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.986382] env[62914]: DEBUG nova.scheduler.client.report [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.003639] env[62914]: DEBUG oslo_vmware.rw_handles [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Completed reading data from the image iterator. {{(pid=62914) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 969.003935] env[62914]: DEBUG oslo_vmware.rw_handles [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5221e259-f9ca-f567-e6f0-5f63c7cd4d9e/disk-0.vmdk. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 969.005618] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e0f99b-d8fb-456b-b270-984fdc81d416 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.013278] env[62914]: DEBUG oslo_vmware.rw_handles [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5221e259-f9ca-f567-e6f0-5f63c7cd4d9e/disk-0.vmdk is in state: ready. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 969.013605] env[62914]: DEBUG oslo_vmware.rw_handles [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5221e259-f9ca-f567-e6f0-5f63c7cd4d9e/disk-0.vmdk. {{(pid=62914) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 969.013716] env[62914]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-9d5cbe6c-cdba-401d-9375-21cf501c1788 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.230248] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ab5d21-2666-948d-79b7-89181a69c92b, 'name': SearchDatastore_Task, 'duration_secs': 0.02143} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.230579] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.230821] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 969.231079] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.231286] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.231531] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 969.231806] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0edaf59-7d1c-4390-a2eb-92d205432c7a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.240410] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 969.240601] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 969.241370] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19126595-113a-42b9-ad08-960e29120701 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.247181] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 969.247181] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]522a1bb2-5f1b-a015-dae1-789fe9625f22" [ 969.247181] env[62914]: _type = "Task" [ 969.247181] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.261677] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]522a1bb2-5f1b-a015-dae1-789fe9625f22, 'name': SearchDatastore_Task, 'duration_secs': 0.012101} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.262506] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8ba320b-9033-4d85-842a-33c63b6d3d0f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.266257] env[62914]: DEBUG oslo_vmware.rw_handles [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5221e259-f9ca-f567-e6f0-5f63c7cd4d9e/disk-0.vmdk. {{(pid=62914) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 969.266497] env[62914]: INFO nova.virt.vmwareapi.images [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Downloaded image file data 67e90822-61d3-4074-a5de-33c47f66782e [ 969.267670] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04435b6-4fe7-4349-950b-1f6558343b37 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.271408] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 969.271408] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]524a1703-6de2-814d-6e18-fdfd090f45fb" [ 969.271408] env[62914]: _type = "Task" [ 969.271408] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.285668] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0db18a47-f262-4bb7-820a-ea5c93f0dd43 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.290735] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]524a1703-6de2-814d-6e18-fdfd090f45fb, 'name': SearchDatastore_Task, 'duration_secs': 0.009078} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.290984] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.291296] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 5e8f452f-68e6-4fe6-9955-20359d90851b/5e8f452f-68e6-4fe6-9955-20359d90851b.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 969.291559] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2d182b1-edb9-46db-af8c-315daa060e68 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.297189] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 969.297189] env[62914]: value = "task-1352849" [ 969.297189] env[62914]: _type = "Task" [ 969.297189] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.305426] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352849, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.318881] env[62914]: INFO nova.virt.vmwareapi.images [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] The imported VM was unregistered [ 969.321577] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Caching image {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 969.321830] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Creating directory with path [datastore2] devstack-image-cache_base/67e90822-61d3-4074-a5de-33c47f66782e {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 969.322197] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95c8d59d-d629-43dd-8342-e20cd729a865 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.340850] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Created directory with path [datastore2] devstack-image-cache_base/67e90822-61d3-4074-a5de-33c47f66782e {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 969.341079] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_b69cb349-e00e-428c-8078-55c1c7577ae8/OSTACK_IMG_b69cb349-e00e-428c-8078-55c1c7577ae8.vmdk to [datastore2] devstack-image-cache_base/67e90822-61d3-4074-a5de-33c47f66782e/67e90822-61d3-4074-a5de-33c47f66782e.vmdk. {{(pid=62914) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 969.341376] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-d5e7f420-e305-4f06-be47-e64a2c3ae9fa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.348145] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 969.348145] env[62914]: value = "task-1352850" [ 969.348145] env[62914]: _type = "Task" [ 969.348145] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.357927] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Releasing lock "refresh_cache-c7cf71d7-b823-471f-949f-9ef2b539d972" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.358144] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updated the network info_cache for instance {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 969.358397] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352850, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.358600] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.358756] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.358922] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.359088] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.360242] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.360242] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.360242] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62914) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 969.360242] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager.update_available_resource {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 969.492692] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.228s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.493324] env[62914]: DEBUG nova.compute.manager [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 969.818699] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352849, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483243} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.819042] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 5e8f452f-68e6-4fe6-9955-20359d90851b/5e8f452f-68e6-4fe6-9955-20359d90851b.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 969.819602] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 969.819602] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-488e2aae-cb0f-4306-9559-617ded0e6ff5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.828946] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 969.828946] env[62914]: value = "task-1352851" [ 969.828946] env[62914]: _type = "Task" [ 969.828946] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.841320] env[62914]: DEBUG nova.compute.manager [req-f24e76bf-6bc6-4d44-8b3f-418ceb4e81db req-74f87e83-34ec-4a4e-a06d-36446207b019 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Received event network-vif-plugged-3e7456bf-2578-404e-98da-f253c33ac065 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.841440] env[62914]: DEBUG oslo_concurrency.lockutils [req-f24e76bf-6bc6-4d44-8b3f-418ceb4e81db req-74f87e83-34ec-4a4e-a06d-36446207b019 service nova] Acquiring lock "1e527707-758c-4006-af83-9c739b9645ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.841673] env[62914]: DEBUG oslo_concurrency.lockutils [req-f24e76bf-6bc6-4d44-8b3f-418ceb4e81db req-74f87e83-34ec-4a4e-a06d-36446207b019 service nova] Lock "1e527707-758c-4006-af83-9c739b9645ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.842200] env[62914]: DEBUG oslo_concurrency.lockutils [req-f24e76bf-6bc6-4d44-8b3f-418ceb4e81db req-74f87e83-34ec-4a4e-a06d-36446207b019 service nova] Lock "1e527707-758c-4006-af83-9c739b9645ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.842364] env[62914]: DEBUG nova.compute.manager [req-f24e76bf-6bc6-4d44-8b3f-418ceb4e81db req-74f87e83-34ec-4a4e-a06d-36446207b019 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] No waiting events found dispatching network-vif-plugged-3e7456bf-2578-404e-98da-f253c33ac065 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 969.842593] env[62914]: WARNING nova.compute.manager [req-f24e76bf-6bc6-4d44-8b3f-418ceb4e81db req-74f87e83-34ec-4a4e-a06d-36446207b019 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Received unexpected event network-vif-plugged-3e7456bf-2578-404e-98da-f253c33ac065 for instance with vm_state active and task_state None. [ 969.846749] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352851, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.861313] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352850, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.864987] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.865344] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.865532] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.865692] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62914) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 969.866728] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b73f0177-2886-4935-94a3-99bea96e3388 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.877728] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01cd985-e2bf-45b4-bac6-baa6164deeb8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.899887] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b4fdea-eba3-4268-b6c3-36eede3bebf8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.911218] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8736368b-7f92-493d-91df-eafdb7eff2df {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.948818] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180776MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62914) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 969.949058] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.949398] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.955839] env[62914]: DEBUG nova.network.neutron [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Successfully updated port: 3e7456bf-2578-404e-98da-f253c33ac065 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 969.998513] env[62914]: DEBUG nova.compute.utils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 970.000584] env[62914]: DEBUG nova.compute.manager [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 970.000818] env[62914]: DEBUG nova.network.neutron [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 970.043359] env[62914]: DEBUG nova.policy [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '795b94761b144a1e99ed79838d9687ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9278818cebd842669b802a4e43e91774', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 970.307099] env[62914]: DEBUG nova.network.neutron [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Successfully created port: 7d0c685f-77d2-4027-a19d-ecc0ac88106f {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 970.341656] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352851, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.182484} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.341990] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 970.342835] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-220d9681-ff87-4433-953d-f197ed851b52 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.372831] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 5e8f452f-68e6-4fe6-9955-20359d90851b/5e8f452f-68e6-4fe6-9955-20359d90851b.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 970.374054] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c76cbe4f-6c40-45a2-9c04-c5239c03fdd5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.393739] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352850, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.403112] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 970.403112] env[62914]: value = "task-1352852" [ 970.403112] env[62914]: _type = "Task" [ 970.403112] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.414027] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352852, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.462446] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.462658] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.462734] env[62914]: DEBUG nova.network.neutron [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 970.505631] env[62914]: DEBUG nova.compute.manager [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 970.862356] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352850, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.919536] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.992033] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 1e527707-758c-4006-af83-9c739b9645ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.992033] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 412c706a-31ce-4a33-914d-5762b02b4c63 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.992296] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance c7cf71d7-b823-471f-949f-9ef2b539d972 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.992350] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 2575bc0d-cb47-4dab-a219-f84254bda47d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.992451] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.992580] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance a2a13ec9-30f3-4c04-b10b-f9a9c01534bd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.992696] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 5e8f452f-68e6-4fe6-9955-20359d90851b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.992807] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.993031] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 970.993208] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 971.006490] env[62914]: WARNING nova.network.neutron [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] f59e1bed-468c-492e-a501-34925e946a66 already exists in list: networks containing: ['f59e1bed-468c-492e-a501-34925e946a66']. ignoring it [ 971.131486] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ab7396-a73b-4e95-8404-6f476a9dab43 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.143478] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdfc68e1-8132-4ef8-93ce-5250088e453f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.189189] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8b4839-aa97-41f9-91fb-fb927fb2c9e7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.199907] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c18dce2-702e-4a10-96ca-ad7f37323620 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.219058] env[62914]: DEBUG nova.compute.provider_tree [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.346634] env[62914]: DEBUG nova.network.neutron [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updating instance_info_cache with network_info: [{"id": "615c8b1c-438d-43fb-999b-89172ade3f61", "address": "fa:16:3e:9d:b8:15", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap615c8b1c-43", "ovs_interfaceid": "615c8b1c-438d-43fb-999b-89172ade3f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3e7456bf-2578-404e-98da-f253c33ac065", "address": "fa:16:3e:8d:7a:56", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e7456bf-25", "ovs_interfaceid": "3e7456bf-2578-404e-98da-f253c33ac065", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.363969] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352850, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.418514] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.519242] env[62914]: DEBUG nova.compute.manager [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 971.547763] env[62914]: DEBUG nova.virt.hardware [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 971.548084] env[62914]: DEBUG nova.virt.hardware [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 971.548351] env[62914]: DEBUG nova.virt.hardware [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 971.548564] env[62914]: DEBUG nova.virt.hardware [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 971.548786] env[62914]: DEBUG nova.virt.hardware [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 971.549037] env[62914]: DEBUG nova.virt.hardware [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 971.549382] env[62914]: DEBUG nova.virt.hardware [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 971.549582] env[62914]: DEBUG nova.virt.hardware [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 971.549812] env[62914]: DEBUG nova.virt.hardware [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 971.550052] env[62914]: DEBUG nova.virt.hardware [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 971.550274] env[62914]: DEBUG nova.virt.hardware [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.551211] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddeb3793-323c-4939-966e-e87ea13d507e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.562356] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7efdd569-b777-4738-abd3-bbbb0ccfcd72 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.722619] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.849996] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.851037] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.851037] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.851930] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54544cb2-746d-4fb8-a6ab-92d7380d9d96 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.873883] env[62914]: DEBUG nova.virt.hardware [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 971.874446] env[62914]: DEBUG nova.virt.hardware [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 971.874446] env[62914]: DEBUG nova.virt.hardware [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 971.874593] env[62914]: DEBUG nova.virt.hardware [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 971.874757] env[62914]: DEBUG nova.virt.hardware [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 971.874877] env[62914]: DEBUG nova.virt.hardware [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 971.875100] env[62914]: DEBUG nova.virt.hardware [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 971.875273] env[62914]: DEBUG nova.virt.hardware [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 971.875442] env[62914]: DEBUG nova.virt.hardware [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 971.875607] env[62914]: DEBUG nova.virt.hardware [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 971.875783] env[62914]: DEBUG nova.virt.hardware [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.882450] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Reconfiguring VM to attach interface {{(pid=62914) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 971.885926] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29f48b0f-8a87-4882-baf3-bea29206704c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.898597] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352850, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.908834] env[62914]: DEBUG oslo_vmware.api [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 971.908834] env[62914]: value = "task-1352853" [ 971.908834] env[62914]: _type = "Task" [ 971.908834] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.916441] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352852, 'name': ReconfigVM_Task, 'duration_secs': 1.495091} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.917193] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 5e8f452f-68e6-4fe6-9955-20359d90851b/5e8f452f-68e6-4fe6-9955-20359d90851b.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 971.917834] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-08539aff-3f71-466d-9061-3cc0a7396f42 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.922529] env[62914]: DEBUG oslo_vmware.api [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352853, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.929167] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 971.929167] env[62914]: value = "task-1352854" [ 971.929167] env[62914]: _type = "Task" [ 971.929167] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.942614] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352854, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.984926] env[62914]: DEBUG nova.network.neutron [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Successfully updated port: 7d0c685f-77d2-4027-a19d-ecc0ac88106f {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 972.003402] env[62914]: DEBUG nova.compute.manager [req-8b814cc6-ea43-4f58-aa8c-c611f6c1d725 req-29ff87f1-cb7d-4dd6-b188-6003ef268989 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Received event network-changed-3e7456bf-2578-404e-98da-f253c33ac065 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.003644] env[62914]: DEBUG nova.compute.manager [req-8b814cc6-ea43-4f58-aa8c-c611f6c1d725 req-29ff87f1-cb7d-4dd6-b188-6003ef268989 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Refreshing instance network info cache due to event network-changed-3e7456bf-2578-404e-98da-f253c33ac065. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 972.003920] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b814cc6-ea43-4f58-aa8c-c611f6c1d725 req-29ff87f1-cb7d-4dd6-b188-6003ef268989 service nova] Acquiring lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.004885] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b814cc6-ea43-4f58-aa8c-c611f6c1d725 req-29ff87f1-cb7d-4dd6-b188-6003ef268989 service nova] Acquired lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.004885] env[62914]: DEBUG nova.network.neutron [req-8b814cc6-ea43-4f58-aa8c-c611f6c1d725 req-29ff87f1-cb7d-4dd6-b188-6003ef268989 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Refreshing network info cache for port 3e7456bf-2578-404e-98da-f253c33ac065 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 972.227985] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 972.228228] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.279s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.361055] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352850, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.981656} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.361410] env[62914]: INFO nova.virt.vmwareapi.ds_util [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_b69cb349-e00e-428c-8078-55c1c7577ae8/OSTACK_IMG_b69cb349-e00e-428c-8078-55c1c7577ae8.vmdk to [datastore2] devstack-image-cache_base/67e90822-61d3-4074-a5de-33c47f66782e/67e90822-61d3-4074-a5de-33c47f66782e.vmdk. [ 972.361532] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Cleaning up location [datastore2] OSTACK_IMG_b69cb349-e00e-428c-8078-55c1c7577ae8 {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 972.361635] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_b69cb349-e00e-428c-8078-55c1c7577ae8 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.361891] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7e20093d-e5a8-4217-873d-f76a818c153a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.368154] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 972.368154] env[62914]: value = "task-1352855" [ 972.368154] env[62914]: _type = "Task" [ 972.368154] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.375950] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352855, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.417254] env[62914]: DEBUG oslo_vmware.api [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352853, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.437717] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352854, 'name': Rename_Task, 'duration_secs': 0.136459} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.438118] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.438429] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d770a226-7bf4-4765-9582-4391631c18a0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.444468] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 972.444468] env[62914]: value = "task-1352856" [ 972.444468] env[62914]: _type = "Task" [ 972.444468] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.457386] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352856, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.488643] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "refresh_cache-6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.488909] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "refresh_cache-6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.489168] env[62914]: DEBUG nova.network.neutron [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 972.715158] env[62914]: DEBUG nova.network.neutron [req-8b814cc6-ea43-4f58-aa8c-c611f6c1d725 req-29ff87f1-cb7d-4dd6-b188-6003ef268989 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updated VIF entry in instance network info cache for port 3e7456bf-2578-404e-98da-f253c33ac065. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 972.715615] env[62914]: DEBUG nova.network.neutron [req-8b814cc6-ea43-4f58-aa8c-c611f6c1d725 req-29ff87f1-cb7d-4dd6-b188-6003ef268989 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updating instance_info_cache with network_info: [{"id": "615c8b1c-438d-43fb-999b-89172ade3f61", "address": "fa:16:3e:9d:b8:15", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap615c8b1c-43", "ovs_interfaceid": "615c8b1c-438d-43fb-999b-89172ade3f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3e7456bf-2578-404e-98da-f253c33ac065", "address": "fa:16:3e:8d:7a:56", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e7456bf-25", "ovs_interfaceid": "3e7456bf-2578-404e-98da-f253c33ac065", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.878374] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352855, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.032102} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.878570] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.878755] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "[datastore2] devstack-image-cache_base/67e90822-61d3-4074-a5de-33c47f66782e/67e90822-61d3-4074-a5de-33c47f66782e.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.879019] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/67e90822-61d3-4074-a5de-33c47f66782e/67e90822-61d3-4074-a5de-33c47f66782e.vmdk to [datastore2] a2a13ec9-30f3-4c04-b10b-f9a9c01534bd/a2a13ec9-30f3-4c04-b10b-f9a9c01534bd.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 972.879337] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b8bcf624-1614-46e8-8624-4b1998646d17 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.885919] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 972.885919] env[62914]: value = "task-1352857" [ 972.885919] env[62914]: _type = "Task" [ 972.885919] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.893824] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352857, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.918730] env[62914]: DEBUG oslo_vmware.api [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352853, 'name': ReconfigVM_Task, 'duration_secs': 0.768496} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.919250] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.919466] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Reconfigured VM to attach interface {{(pid=62914) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 972.955059] env[62914]: DEBUG oslo_vmware.api [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352856, 'name': PowerOnVM_Task, 'duration_secs': 0.486355} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.955356] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 972.955564] env[62914]: INFO nova.compute.manager [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Took 7.80 seconds to spawn the instance on the hypervisor. [ 972.955742] env[62914]: DEBUG nova.compute.manager [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 972.956740] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27192543-661d-4daf-9b10-215ca1352a89 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.024651] env[62914]: DEBUG nova.network.neutron [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 973.218948] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b814cc6-ea43-4f58-aa8c-c611f6c1d725 req-29ff87f1-cb7d-4dd6-b188-6003ef268989 service nova] Releasing lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.219292] env[62914]: DEBUG nova.compute.manager [req-8b814cc6-ea43-4f58-aa8c-c611f6c1d725 req-29ff87f1-cb7d-4dd6-b188-6003ef268989 service nova] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Received event network-vif-plugged-7d0c685f-77d2-4027-a19d-ecc0ac88106f {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 973.219620] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b814cc6-ea43-4f58-aa8c-c611f6c1d725 req-29ff87f1-cb7d-4dd6-b188-6003ef268989 service nova] Acquiring lock "6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.219926] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b814cc6-ea43-4f58-aa8c-c611f6c1d725 req-29ff87f1-cb7d-4dd6-b188-6003ef268989 service nova] Lock "6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.220294] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b814cc6-ea43-4f58-aa8c-c611f6c1d725 req-29ff87f1-cb7d-4dd6-b188-6003ef268989 service nova] Lock "6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.220598] env[62914]: DEBUG nova.compute.manager [req-8b814cc6-ea43-4f58-aa8c-c611f6c1d725 req-29ff87f1-cb7d-4dd6-b188-6003ef268989 service nova] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] No waiting events found dispatching network-vif-plugged-7d0c685f-77d2-4027-a19d-ecc0ac88106f {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 973.220812] env[62914]: WARNING nova.compute.manager [req-8b814cc6-ea43-4f58-aa8c-c611f6c1d725 req-29ff87f1-cb7d-4dd6-b188-6003ef268989 service nova] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Received unexpected event network-vif-plugged-7d0c685f-77d2-4027-a19d-ecc0ac88106f for instance with vm_state building and task_state spawning. [ 973.231375] env[62914]: DEBUG nova.network.neutron [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Updating instance_info_cache with network_info: [{"id": "7d0c685f-77d2-4027-a19d-ecc0ac88106f", "address": "fa:16:3e:d1:fe:84", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d0c685f-77", "ovs_interfaceid": "7d0c685f-77d2-4027-a19d-ecc0ac88106f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.398186] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352857, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.424803] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d3f8b42-bd53-4fb6-aae2-c800732b1472 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-1e527707-758c-4006-af83-9c739b9645ed-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.491s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.475356] env[62914]: INFO nova.compute.manager [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Took 13.44 seconds to build instance. [ 973.734769] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "refresh_cache-6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.735069] env[62914]: DEBUG nova.compute.manager [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Instance network_info: |[{"id": "7d0c685f-77d2-4027-a19d-ecc0ac88106f", "address": "fa:16:3e:d1:fe:84", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d0c685f-77", "ovs_interfaceid": "7d0c685f-77d2-4027-a19d-ecc0ac88106f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 973.735542] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:fe:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '32faf59b-014c-4f1f-8331-40df95bf741f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d0c685f-77d2-4027-a19d-ecc0ac88106f', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 973.743891] env[62914]: DEBUG oslo.service.loopingcall [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.744150] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 973.744390] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-81878d0a-e305-40df-889c-c2b7441dfc00 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.766694] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 973.766694] env[62914]: value = "task-1352858" [ 973.766694] env[62914]: _type = "Task" [ 973.766694] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.777290] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352858, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.899205] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352857, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.977774] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9696ea48-d85c-4240-a2f4-abd2f9d19862 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "5e8f452f-68e6-4fe6-9955-20359d90851b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.950s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.042654] env[62914]: DEBUG nova.compute.manager [req-b5cb7e2e-9468-4f5b-a153-bbb9d4cb92b2 req-13c67c9c-5bfd-40f9-9150-e8dd22743f99 service nova] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Received event network-changed-7d0c685f-77d2-4027-a19d-ecc0ac88106f {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 974.043060] env[62914]: DEBUG nova.compute.manager [req-b5cb7e2e-9468-4f5b-a153-bbb9d4cb92b2 req-13c67c9c-5bfd-40f9-9150-e8dd22743f99 service nova] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Refreshing instance network info cache due to event network-changed-7d0c685f-77d2-4027-a19d-ecc0ac88106f. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 974.043383] env[62914]: DEBUG oslo_concurrency.lockutils [req-b5cb7e2e-9468-4f5b-a153-bbb9d4cb92b2 req-13c67c9c-5bfd-40f9-9150-e8dd22743f99 service nova] Acquiring lock "refresh_cache-6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.043612] env[62914]: DEBUG oslo_concurrency.lockutils [req-b5cb7e2e-9468-4f5b-a153-bbb9d4cb92b2 req-13c67c9c-5bfd-40f9-9150-e8dd22743f99 service nova] Acquired lock "refresh_cache-6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.043865] env[62914]: DEBUG nova.network.neutron [req-b5cb7e2e-9468-4f5b-a153-bbb9d4cb92b2 req-13c67c9c-5bfd-40f9-9150-e8dd22743f99 service nova] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Refreshing network info cache for port 7d0c685f-77d2-4027-a19d-ecc0ac88106f {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 974.280028] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352858, 'name': CreateVM_Task, 'duration_secs': 0.391641} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.280028] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 974.280028] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.280028] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.280562] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 974.280562] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-780465ed-20c4-41ed-a423-78a4a290f28f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.288144] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 974.288144] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5255c1f6-3ddf-2be8-44e3-bb6f7c2cb103" [ 974.288144] env[62914]: _type = "Task" [ 974.288144] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.299760] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5255c1f6-3ddf-2be8-44e3-bb6f7c2cb103, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.399262] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352857, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.589240] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b15a211-5793-4c7a-8467-98d8aece4069 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "5e8f452f-68e6-4fe6-9955-20359d90851b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.591814] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b15a211-5793-4c7a-8467-98d8aece4069 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "5e8f452f-68e6-4fe6-9955-20359d90851b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.591814] env[62914]: DEBUG nova.compute.manager [None req-3b15a211-5793-4c7a-8467-98d8aece4069 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 974.591814] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161c67b3-ba47-464f-8a02-861f6f321dd6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.602886] env[62914]: DEBUG nova.compute.manager [None req-3b15a211-5793-4c7a-8467-98d8aece4069 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62914) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 974.602886] env[62914]: DEBUG nova.objects.instance [None req-3b15a211-5793-4c7a-8467-98d8aece4069 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lazy-loading 'flavor' on Instance uuid 5e8f452f-68e6-4fe6-9955-20359d90851b {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.762498] env[62914]: DEBUG nova.network.neutron [req-b5cb7e2e-9468-4f5b-a153-bbb9d4cb92b2 req-13c67c9c-5bfd-40f9-9150-e8dd22743f99 service nova] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Updated VIF entry in instance network info cache for port 7d0c685f-77d2-4027-a19d-ecc0ac88106f. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 974.762897] env[62914]: DEBUG nova.network.neutron [req-b5cb7e2e-9468-4f5b-a153-bbb9d4cb92b2 req-13c67c9c-5bfd-40f9-9150-e8dd22743f99 service nova] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Updating instance_info_cache with network_info: [{"id": "7d0c685f-77d2-4027-a19d-ecc0ac88106f", "address": "fa:16:3e:d1:fe:84", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d0c685f-77", "ovs_interfaceid": "7d0c685f-77d2-4027-a19d-ecc0ac88106f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.801312] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5255c1f6-3ddf-2be8-44e3-bb6f7c2cb103, 'name': SearchDatastore_Task, 'duration_secs': 0.014445} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.801693] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.801957] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 974.802272] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.802471] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.802694] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 974.802999] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7741bcb5-7347-47d8-a6ac-72fb611fee6e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.814386] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 974.814579] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 974.815362] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e340b50-40c3-406b-a08a-321bc8f48bf7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.822301] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 974.822301] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5204bafb-3c23-e79b-4aef-1126901d2fb4" [ 974.822301] env[62914]: _type = "Task" [ 974.822301] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.831358] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5204bafb-3c23-e79b-4aef-1126901d2fb4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.899741] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352857, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.046815] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "interface-1e527707-758c-4006-af83-9c739b9645ed-285c9a22-ca86-4373-8394-fab932eddd6f" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.046815] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-1e527707-758c-4006-af83-9c739b9645ed-285c9a22-ca86-4373-8394-fab932eddd6f" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.047151] env[62914]: DEBUG nova.objects.instance [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'flavor' on Instance uuid 1e527707-758c-4006-af83-9c739b9645ed {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.108377] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b15a211-5793-4c7a-8467-98d8aece4069 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 975.108664] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-268725d3-643c-434f-9072-a4086754a06a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.117456] env[62914]: DEBUG oslo_vmware.api [None req-3b15a211-5793-4c7a-8467-98d8aece4069 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 975.117456] env[62914]: value = "task-1352859" [ 975.117456] env[62914]: _type = "Task" [ 975.117456] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.125939] env[62914]: DEBUG oslo_vmware.api [None req-3b15a211-5793-4c7a-8467-98d8aece4069 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352859, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.265791] env[62914]: DEBUG oslo_concurrency.lockutils [req-b5cb7e2e-9468-4f5b-a153-bbb9d4cb92b2 req-13c67c9c-5bfd-40f9-9150-e8dd22743f99 service nova] Releasing lock "refresh_cache-6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.332924] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5204bafb-3c23-e79b-4aef-1126901d2fb4, 'name': SearchDatastore_Task, 'duration_secs': 0.01298} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.333831] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c5b6615-2a38-4666-8198-7ec334d0e2d3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.338828] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 975.338828] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]528a1605-c55f-9c9b-f7a7-8b979b9ebb5d" [ 975.338828] env[62914]: _type = "Task" [ 975.338828] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.346530] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]528a1605-c55f-9c9b-f7a7-8b979b9ebb5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.399008] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352857, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.247625} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.399294] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/67e90822-61d3-4074-a5de-33c47f66782e/67e90822-61d3-4074-a5de-33c47f66782e.vmdk to [datastore2] a2a13ec9-30f3-4c04-b10b-f9a9c01534bd/a2a13ec9-30f3-4c04-b10b-f9a9c01534bd.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 975.400149] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c211c7e5-41b9-445d-a98a-918a6b668bc9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.422093] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] a2a13ec9-30f3-4c04-b10b-f9a9c01534bd/a2a13ec9-30f3-4c04-b10b-f9a9c01534bd.vmdk or device None with type streamOptimized {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.422423] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29e43666-e634-48b2-b624-e8798f37202c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.441744] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 975.441744] env[62914]: value = "task-1352860" [ 975.441744] env[62914]: _type = "Task" [ 975.441744] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.449511] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352860, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.628863] env[62914]: DEBUG oslo_vmware.api [None req-3b15a211-5793-4c7a-8467-98d8aece4069 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352859, 'name': PowerOffVM_Task, 'duration_secs': 0.272131} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.629153] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b15a211-5793-4c7a-8467-98d8aece4069 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 975.629333] env[62914]: DEBUG nova.compute.manager [None req-3b15a211-5793-4c7a-8467-98d8aece4069 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.630095] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd7e28ff-cfa9-4f08-917c-4f3c09535400 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.644137] env[62914]: DEBUG nova.objects.instance [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'pci_requests' on Instance uuid 1e527707-758c-4006-af83-9c739b9645ed {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.848902] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]528a1605-c55f-9c9b-f7a7-8b979b9ebb5d, 'name': SearchDatastore_Task, 'duration_secs': 0.009733} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.849208] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.849465] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509/6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 975.849720] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b2e1adf-9d8e-420f-97c8-c9c47265b84b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.855765] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 975.855765] env[62914]: value = "task-1352861" [ 975.855765] env[62914]: _type = "Task" [ 975.855765] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.863342] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352861, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.951349] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352860, 'name': ReconfigVM_Task, 'duration_secs': 0.307992} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.951642] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Reconfigured VM instance instance-0000005e to attach disk [datastore2] a2a13ec9-30f3-4c04-b10b-f9a9c01534bd/a2a13ec9-30f3-4c04-b10b-f9a9c01534bd.vmdk or device None with type streamOptimized {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 975.952461] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8326b959-6a61-4ac1-ad8d-7de0c552d549 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.958312] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 975.958312] env[62914]: value = "task-1352862" [ 975.958312] env[62914]: _type = "Task" [ 975.958312] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.966054] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352862, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.141797] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3b15a211-5793-4c7a-8467-98d8aece4069 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "5e8f452f-68e6-4fe6-9955-20359d90851b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.552s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.146754] env[62914]: DEBUG nova.objects.base [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Object Instance<1e527707-758c-4006-af83-9c739b9645ed> lazy-loaded attributes: flavor,pci_requests {{(pid=62914) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 976.146961] env[62914]: DEBUG nova.network.neutron [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 976.219573] env[62914]: DEBUG nova.policy [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80089586cfda4cd2bdfc24d9f556239c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09d51090ac5d41fc966c7e132f288824', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 976.365686] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352861, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467602} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.366067] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509/6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 976.366149] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 976.366401] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-70cf615c-95d7-4125-8e29-09214814837d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.372319] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 976.372319] env[62914]: value = "task-1352863" [ 976.372319] env[62914]: _type = "Task" [ 976.372319] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.379970] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352863, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.467956] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352862, 'name': Rename_Task, 'duration_secs': 0.18113} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.468802] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 976.468802] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a397f50d-021e-4e16-ac87-d4a2d556d1a0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.474670] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 976.474670] env[62914]: value = "task-1352864" [ 976.474670] env[62914]: _type = "Task" [ 976.474670] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.482019] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352864, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.883420] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352863, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064906} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.883647] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 976.884443] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ea603a-0d03-4eb2-a66f-c40664bef5b6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.907516] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509/6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.907801] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85f112ec-121d-4558-b0c6-560d4d16102d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.926908] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 976.926908] env[62914]: value = "task-1352865" [ 976.926908] env[62914]: _type = "Task" [ 976.926908] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.935029] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352865, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.984791] env[62914]: DEBUG oslo_vmware.api [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352864, 'name': PowerOnVM_Task, 'duration_secs': 0.444759} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.985088] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 976.985291] env[62914]: INFO nova.compute.manager [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Took 14.05 seconds to spawn the instance on the hypervisor. [ 976.985469] env[62914]: DEBUG nova.compute.manager [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 976.986257] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698ed801-5d93-452b-a116-162e793f37fa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.988845] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "5e8f452f-68e6-4fe6-9955-20359d90851b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.989079] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "5e8f452f-68e6-4fe6-9955-20359d90851b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.989297] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "5e8f452f-68e6-4fe6-9955-20359d90851b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.989460] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "5e8f452f-68e6-4fe6-9955-20359d90851b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.989629] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "5e8f452f-68e6-4fe6-9955-20359d90851b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.991979] env[62914]: INFO nova.compute.manager [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Terminating instance [ 976.997345] env[62914]: DEBUG nova.compute.manager [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 976.997534] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 976.998255] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b92b8b7-40e4-4bf2-82c5-80cae23fa76d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.004737] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 977.004950] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e733dc5b-ce4c-4b71-963f-44ed11d7cf57 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.069145] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 977.069407] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 977.069620] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleting the datastore file [datastore2] 5e8f452f-68e6-4fe6-9955-20359d90851b {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 977.069906] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a4e1f14f-78f6-4953-8bd6-0c2ef8ddf5ad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.077183] env[62914]: DEBUG oslo_vmware.api [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 977.077183] env[62914]: value = "task-1352867" [ 977.077183] env[62914]: _type = "Task" [ 977.077183] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.084731] env[62914]: DEBUG oslo_vmware.api [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352867, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.437725] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352865, 'name': ReconfigVM_Task, 'duration_secs': 0.32855} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.438176] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509/6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 977.438682] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d0b9cf31-f127-4be0-a07d-4013f0151314 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.445730] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 977.445730] env[62914]: value = "task-1352868" [ 977.445730] env[62914]: _type = "Task" [ 977.445730] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.453108] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352868, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.508750] env[62914]: INFO nova.compute.manager [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Took 20.53 seconds to build instance. [ 977.587193] env[62914]: DEBUG oslo_vmware.api [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352867, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.699478] env[62914]: DEBUG nova.compute.manager [req-57bcef4d-e497-47a6-b771-08a7f7bdbfab req-da47d9b2-d998-4658-93c1-22d6e6ed2441 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Received event network-vif-plugged-285c9a22-ca86-4373-8394-fab932eddd6f {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.699725] env[62914]: DEBUG oslo_concurrency.lockutils [req-57bcef4d-e497-47a6-b771-08a7f7bdbfab req-da47d9b2-d998-4658-93c1-22d6e6ed2441 service nova] Acquiring lock "1e527707-758c-4006-af83-9c739b9645ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.699944] env[62914]: DEBUG oslo_concurrency.lockutils [req-57bcef4d-e497-47a6-b771-08a7f7bdbfab req-da47d9b2-d998-4658-93c1-22d6e6ed2441 service nova] Lock "1e527707-758c-4006-af83-9c739b9645ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.700223] env[62914]: DEBUG oslo_concurrency.lockutils [req-57bcef4d-e497-47a6-b771-08a7f7bdbfab req-da47d9b2-d998-4658-93c1-22d6e6ed2441 service nova] Lock "1e527707-758c-4006-af83-9c739b9645ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.700426] env[62914]: DEBUG nova.compute.manager [req-57bcef4d-e497-47a6-b771-08a7f7bdbfab req-da47d9b2-d998-4658-93c1-22d6e6ed2441 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] No waiting events found dispatching network-vif-plugged-285c9a22-ca86-4373-8394-fab932eddd6f {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 977.700607] env[62914]: WARNING nova.compute.manager [req-57bcef4d-e497-47a6-b771-08a7f7bdbfab req-da47d9b2-d998-4658-93c1-22d6e6ed2441 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Received unexpected event network-vif-plugged-285c9a22-ca86-4373-8394-fab932eddd6f for instance with vm_state active and task_state None. [ 977.724125] env[62914]: DEBUG nova.network.neutron [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Successfully updated port: 285c9a22-ca86-4373-8394-fab932eddd6f {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 977.957067] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352868, 'name': Rename_Task, 'duration_secs': 0.178708} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.957067] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 977.957356] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbf05ee3-08e4-42a0-820f-ddaeeb92d5df {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.963360] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 977.963360] env[62914]: value = "task-1352869" [ 977.963360] env[62914]: _type = "Task" [ 977.963360] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.978846] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352869, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.013296] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d539dfc3-2294-479e-9191-88d8d57ad371 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "a2a13ec9-30f3-4c04-b10b-f9a9c01534bd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.046s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.086793] env[62914]: DEBUG oslo_vmware.api [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352867, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.539576} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.087107] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 978.087304] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 978.087482] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 978.087657] env[62914]: INFO nova.compute.manager [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Took 1.09 seconds to destroy the instance on the hypervisor. [ 978.087917] env[62914]: DEBUG oslo.service.loopingcall [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.088107] env[62914]: DEBUG nova.compute.manager [-] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 978.088203] env[62914]: DEBUG nova.network.neutron [-] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 978.227129] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.227129] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.227433] env[62914]: DEBUG nova.network.neutron [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 978.343306] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "a2a13ec9-30f3-4c04-b10b-f9a9c01534bd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.343619] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "a2a13ec9-30f3-4c04-b10b-f9a9c01534bd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.343839] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "a2a13ec9-30f3-4c04-b10b-f9a9c01534bd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.344038] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "a2a13ec9-30f3-4c04-b10b-f9a9c01534bd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.344213] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "a2a13ec9-30f3-4c04-b10b-f9a9c01534bd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.346798] env[62914]: INFO nova.compute.manager [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Terminating instance [ 978.348664] env[62914]: DEBUG nova.compute.manager [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 978.348894] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 978.349762] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d677e442-334f-4d75-8294-2a1aaed31fcf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.358075] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 978.358335] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dda1f5f4-5ff1-48ae-830e-c91fb64b270e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.364680] env[62914]: DEBUG oslo_vmware.api [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 978.364680] env[62914]: value = "task-1352870" [ 978.364680] env[62914]: _type = "Task" [ 978.364680] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.372487] env[62914]: DEBUG oslo_vmware.api [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352870, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.473655] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352869, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.772981] env[62914]: WARNING nova.network.neutron [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] f59e1bed-468c-492e-a501-34925e946a66 already exists in list: networks containing: ['f59e1bed-468c-492e-a501-34925e946a66']. ignoring it [ 978.773234] env[62914]: WARNING nova.network.neutron [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] f59e1bed-468c-492e-a501-34925e946a66 already exists in list: networks containing: ['f59e1bed-468c-492e-a501-34925e946a66']. ignoring it [ 978.823801] env[62914]: DEBUG nova.network.neutron [-] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.875497] env[62914]: DEBUG oslo_vmware.api [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352870, 'name': PowerOffVM_Task, 'duration_secs': 0.224424} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.876131] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 978.876131] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 978.876318] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-752da429-0c4a-4cd3-acdb-1f26db5d31a4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.947678] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 978.947678] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 978.947678] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleting the datastore file [datastore2] a2a13ec9-30f3-4c04-b10b-f9a9c01534bd {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 978.949973] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7a08efa-fa0c-4199-a4ce-a85b63cdde5f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.957257] env[62914]: DEBUG oslo_vmware.api [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 978.957257] env[62914]: value = "task-1352872" [ 978.957257] env[62914]: _type = "Task" [ 978.957257] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.965271] env[62914]: DEBUG oslo_vmware.api [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352872, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.973660] env[62914]: DEBUG oslo_vmware.api [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352869, 'name': PowerOnVM_Task, 'duration_secs': 0.515084} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.975911] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 978.976157] env[62914]: INFO nova.compute.manager [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Took 7.46 seconds to spawn the instance on the hypervisor. [ 978.976346] env[62914]: DEBUG nova.compute.manager [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.977126] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a2531a-7d59-4d7e-8e28-e8e38b2302d8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.203619] env[62914]: DEBUG nova.network.neutron [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updating instance_info_cache with network_info: [{"id": "615c8b1c-438d-43fb-999b-89172ade3f61", "address": "fa:16:3e:9d:b8:15", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap615c8b1c-43", "ovs_interfaceid": "615c8b1c-438d-43fb-999b-89172ade3f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3e7456bf-2578-404e-98da-f253c33ac065", "address": "fa:16:3e:8d:7a:56", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e7456bf-25", "ovs_interfaceid": "3e7456bf-2578-404e-98da-f253c33ac065", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "285c9a22-ca86-4373-8394-fab932eddd6f", "address": "fa:16:3e:58:af:43", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap285c9a22-ca", "ovs_interfaceid": "285c9a22-ca86-4373-8394-fab932eddd6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.328481] env[62914]: INFO nova.compute.manager [-] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Took 1.24 seconds to deallocate network for instance. [ 979.467457] env[62914]: DEBUG oslo_vmware.api [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352872, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146247} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.467702] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.467887] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 979.468081] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 979.468260] env[62914]: INFO nova.compute.manager [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Took 1.12 seconds to destroy the instance on the hypervisor. [ 979.468497] env[62914]: DEBUG oslo.service.loopingcall [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.468693] env[62914]: DEBUG nova.compute.manager [-] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 979.468788] env[62914]: DEBUG nova.network.neutron [-] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 979.493067] env[62914]: INFO nova.compute.manager [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Took 12.25 seconds to build instance. [ 979.706991] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.707697] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.707870] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.708745] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5485be29-f023-4bfb-abd6-ca4724630015 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.727426] env[62914]: DEBUG nova.virt.hardware [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.727671] env[62914]: DEBUG nova.virt.hardware [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.727830] env[62914]: DEBUG nova.virt.hardware [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.728026] env[62914]: DEBUG nova.virt.hardware [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.728181] env[62914]: DEBUG nova.virt.hardware [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.728332] env[62914]: DEBUG nova.virt.hardware [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.728537] env[62914]: DEBUG nova.virt.hardware [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.728728] env[62914]: DEBUG nova.virt.hardware [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.728905] env[62914]: DEBUG nova.virt.hardware [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.729089] env[62914]: DEBUG nova.virt.hardware [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.729251] env[62914]: DEBUG nova.virt.hardware [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.736235] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Reconfiguring VM to attach interface {{(pid=62914) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 979.737152] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4dda939-281b-4001-bf89-c855a47ae2b6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.750526] env[62914]: DEBUG nova.compute.manager [req-9693719d-3615-4fb8-b368-06b4cf01678e req-4c6f7a45-5b62-4d5c-bf52-154842b9777a service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Received event network-changed-285c9a22-ca86-4373-8394-fab932eddd6f {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 979.750666] env[62914]: DEBUG nova.compute.manager [req-9693719d-3615-4fb8-b368-06b4cf01678e req-4c6f7a45-5b62-4d5c-bf52-154842b9777a service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Refreshing instance network info cache due to event network-changed-285c9a22-ca86-4373-8394-fab932eddd6f. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 979.750886] env[62914]: DEBUG oslo_concurrency.lockutils [req-9693719d-3615-4fb8-b368-06b4cf01678e req-4c6f7a45-5b62-4d5c-bf52-154842b9777a service nova] Acquiring lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.751044] env[62914]: DEBUG oslo_concurrency.lockutils [req-9693719d-3615-4fb8-b368-06b4cf01678e req-4c6f7a45-5b62-4d5c-bf52-154842b9777a service nova] Acquired lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.751207] env[62914]: DEBUG nova.network.neutron [req-9693719d-3615-4fb8-b368-06b4cf01678e req-4c6f7a45-5b62-4d5c-bf52-154842b9777a service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Refreshing network info cache for port 285c9a22-ca86-4373-8394-fab932eddd6f {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 979.760932] env[62914]: DEBUG oslo_vmware.api [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 979.760932] env[62914]: value = "task-1352873" [ 979.760932] env[62914]: _type = "Task" [ 979.760932] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.770562] env[62914]: DEBUG oslo_vmware.api [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352873, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.835064] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.835372] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.835611] env[62914]: DEBUG nova.objects.instance [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lazy-loading 'resources' on Instance uuid 5e8f452f-68e6-4fe6-9955-20359d90851b {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.994890] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a9d50fb8-79a6-4036-b341-eababa2142c4 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.755s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.257795] env[62914]: DEBUG nova.network.neutron [-] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.272304] env[62914]: DEBUG oslo_vmware.api [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352873, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.467209] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7565bc-f5fc-4423-b198-3c5a41aa3f09 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.473478] env[62914]: DEBUG nova.network.neutron [req-9693719d-3615-4fb8-b368-06b4cf01678e req-4c6f7a45-5b62-4d5c-bf52-154842b9777a service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updated VIF entry in instance network info cache for port 285c9a22-ca86-4373-8394-fab932eddd6f. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 980.473881] env[62914]: DEBUG nova.network.neutron [req-9693719d-3615-4fb8-b368-06b4cf01678e req-4c6f7a45-5b62-4d5c-bf52-154842b9777a service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updating instance_info_cache with network_info: [{"id": "615c8b1c-438d-43fb-999b-89172ade3f61", "address": "fa:16:3e:9d:b8:15", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap615c8b1c-43", "ovs_interfaceid": "615c8b1c-438d-43fb-999b-89172ade3f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3e7456bf-2578-404e-98da-f253c33ac065", "address": "fa:16:3e:8d:7a:56", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e7456bf-25", "ovs_interfaceid": "3e7456bf-2578-404e-98da-f253c33ac065", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "285c9a22-ca86-4373-8394-fab932eddd6f", "address": "fa:16:3e:58:af:43", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap285c9a22-ca", "ovs_interfaceid": "285c9a22-ca86-4373-8394-fab932eddd6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.477902] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2316ee07-3a87-489b-b811-80952035c1ca {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.509997] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2054d38-a2bb-4707-bd9f-bc686aa25daf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.517535] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ec7b02-5024-4cbd-8131-1010a3997e44 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.536797] env[62914]: DEBUG nova.compute.provider_tree [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.763463] env[62914]: INFO nova.compute.manager [-] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Took 1.29 seconds to deallocate network for instance. [ 980.777328] env[62914]: DEBUG oslo_vmware.api [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352873, 'name': ReconfigVM_Task, 'duration_secs': 0.586971} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.777847] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.778460] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Reconfigured VM to attach interface {{(pid=62914) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 980.978780] env[62914]: DEBUG oslo_concurrency.lockutils [req-9693719d-3615-4fb8-b368-06b4cf01678e req-4c6f7a45-5b62-4d5c-bf52-154842b9777a service nova] Releasing lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.979133] env[62914]: DEBUG nova.compute.manager [req-9693719d-3615-4fb8-b368-06b4cf01678e req-4c6f7a45-5b62-4d5c-bf52-154842b9777a service nova] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Received event network-vif-deleted-90f668f6-799d-4a5a-a911-23dad79f3c4d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.028165] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.028366] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.028628] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.028804] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.028968] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.031317] env[62914]: INFO nova.compute.manager [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Terminating instance [ 981.033216] env[62914]: DEBUG nova.compute.manager [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 981.033415] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 981.034267] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb5a006d-4937-449c-88c8-e50472f59cfb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.039361] env[62914]: DEBUG nova.scheduler.client.report [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 981.044376] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 981.044830] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85c182a4-488b-42f7-a9ec-a1c844e0aae4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.051937] env[62914]: DEBUG oslo_vmware.api [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 981.051937] env[62914]: value = "task-1352874" [ 981.051937] env[62914]: _type = "Task" [ 981.051937] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.062295] env[62914]: DEBUG oslo_vmware.api [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352874, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.274487] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.282802] env[62914]: DEBUG oslo_concurrency.lockutils [None req-8c8e09da-7f7b-4be3-98de-6f1f3b9cbbdf tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-1e527707-758c-4006-af83-9c739b9645ed-285c9a22-ca86-4373-8394-fab932eddd6f" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.236s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.545863] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.710s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.548287] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.274s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.548567] env[62914]: DEBUG nova.objects.instance [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lazy-loading 'resources' on Instance uuid a2a13ec9-30f3-4c04-b10b-f9a9c01534bd {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.563294] env[62914]: DEBUG oslo_vmware.api [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352874, 'name': PowerOffVM_Task, 'duration_secs': 0.204171} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.563733] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 981.563913] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 981.564880] env[62914]: INFO nova.scheduler.client.report [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleted allocations for instance 5e8f452f-68e6-4fe6-9955-20359d90851b [ 981.565762] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-74a5c97c-b6aa-464a-b661-f4bdb9416c55 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.629392] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 981.630068] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 981.630068] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleting the datastore file [datastore1] 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.630226] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d43ed5b-75d2-48b9-864a-59f53e8e5bda {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.638067] env[62914]: DEBUG oslo_vmware.api [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 981.638067] env[62914]: value = "task-1352876" [ 981.638067] env[62914]: _type = "Task" [ 981.638067] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.648513] env[62914]: DEBUG oslo_vmware.api [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352876, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.765921] env[62914]: DEBUG nova.compute.manager [req-03316f10-579d-4417-af4a-ed8d37a8ade9 req-228c4944-2a5d-4f3e-9115-8d77a2fd2f7d service nova] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Received event network-vif-deleted-c37fe151-4337-4f57-855a-5b250e70ce13 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.073738] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26d556e0-b73b-41c5-9443-769564b9a67b tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "5e8f452f-68e6-4fe6-9955-20359d90851b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.084s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.145279] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4341006-9bca-464b-a20a-95f6db5c83fb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.149498] env[62914]: DEBUG oslo_vmware.api [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352876, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133728} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.150079] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 982.150287] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 982.150494] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 982.150680] env[62914]: INFO nova.compute.manager [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Took 1.12 seconds to destroy the instance on the hypervisor. [ 982.150913] env[62914]: DEBUG oslo.service.loopingcall [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.151427] env[62914]: DEBUG nova.compute.manager [-] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 982.152026] env[62914]: DEBUG nova.network.neutron [-] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 982.155867] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9f0cf6-512b-4850-91c8-636e1d3c81ea {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.185739] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8725b67d-2085-401d-b106-db8d6d9f88a0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.193456] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff94f228-a78e-419a-8f4c-fb9962f0bb6a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.206244] env[62914]: DEBUG nova.compute.provider_tree [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.710141] env[62914]: DEBUG nova.scheduler.client.report [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.971833] env[62914]: DEBUG oslo_concurrency.lockutils [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "interface-1e527707-758c-4006-af83-9c739b9645ed-3e7456bf-2578-404e-98da-f253c33ac065" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.972146] env[62914]: DEBUG oslo_concurrency.lockutils [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-1e527707-758c-4006-af83-9c739b9645ed-3e7456bf-2578-404e-98da-f253c33ac065" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.982482] env[62914]: DEBUG nova.network.neutron [-] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.195231] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "84a7822d-d46d-4147-ad4c-b9402aa072f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.195463] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "84a7822d-d46d-4147-ad4c-b9402aa072f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.215078] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.667s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.230480] env[62914]: INFO nova.scheduler.client.report [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleted allocations for instance a2a13ec9-30f3-4c04-b10b-f9a9c01534bd [ 983.475455] env[62914]: DEBUG oslo_concurrency.lockutils [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.475755] env[62914]: DEBUG oslo_concurrency.lockutils [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.476798] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d49b6ebd-0c93-4e08-a520-045d5aa72ce3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.484177] env[62914]: INFO nova.compute.manager [-] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Took 1.33 seconds to deallocate network for instance. [ 983.500103] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd539aa-f128-4084-9d9b-30882ee41acc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.530420] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Reconfiguring VM to detach interface {{(pid=62914) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 983.531097] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b69b6ba3-cb5f-424e-b1d4-6b39b5b9fb37 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.551817] env[62914]: DEBUG oslo_vmware.api [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 983.551817] env[62914]: value = "task-1352877" [ 983.551817] env[62914]: _type = "Task" [ 983.551817] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.560982] env[62914]: DEBUG oslo_vmware.api [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352877, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.697994] env[62914]: DEBUG nova.compute.manager [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 983.737532] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f60197a0-ae4e-4d6b-9ae3-6da4e10b40df tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "a2a13ec9-30f3-4c04-b10b-f9a9c01534bd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.394s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.788986] env[62914]: DEBUG nova.compute.manager [req-f5351e24-72a0-4160-86e3-094e8be7b7f7 req-f6673a96-fae4-48e4-a020-dad173086d6e service nova] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Received event network-vif-deleted-7d0c685f-77d2-4027-a19d-ecc0ac88106f {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.007305] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.007681] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.007964] env[62914]: DEBUG nova.objects.instance [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lazy-loading 'resources' on Instance uuid 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.061708] env[62914]: DEBUG oslo_vmware.api [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352877, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.069228] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "412c706a-31ce-4a33-914d-5762b02b4c63" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.069467] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "412c706a-31ce-4a33-914d-5762b02b4c63" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.069674] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "412c706a-31ce-4a33-914d-5762b02b4c63-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.069852] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "412c706a-31ce-4a33-914d-5762b02b4c63-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.070032] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "412c706a-31ce-4a33-914d-5762b02b4c63-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.072012] env[62914]: INFO nova.compute.manager [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Terminating instance [ 984.073983] env[62914]: DEBUG nova.compute.manager [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 984.074193] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 984.074912] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64162aa6-61d1-475e-9d62-76cd0a600609 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.081483] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.081701] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4521f931-aba6-4775-8764-6bcfdcfe7bc5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.087479] env[62914]: DEBUG oslo_vmware.api [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 984.087479] env[62914]: value = "task-1352878" [ 984.087479] env[62914]: _type = "Task" [ 984.087479] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.094659] env[62914]: DEBUG oslo_vmware.api [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352878, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.217660] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.499944] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "054f37ad-4161-4c70-aa1b-b2e69cd1e9f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.500224] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "054f37ad-4161-4c70-aa1b-b2e69cd1e9f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.563591] env[62914]: DEBUG oslo_vmware.api [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352877, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.596683] env[62914]: DEBUG oslo_vmware.api [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352878, 'name': PowerOffVM_Task, 'duration_secs': 0.184844} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.599756] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 984.599937] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 984.600500] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8fe6c2f1-3346-4be3-aa77-ffec7c8142cb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.603149] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "2575bc0d-cb47-4dab-a219-f84254bda47d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.603369] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "2575bc0d-cb47-4dab-a219-f84254bda47d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.610193] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ec1d57-cb7e-41d1-9d6e-21fec30c795a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.617512] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de5dec4b-43b5-458d-a0eb-187ce4350d41 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.647819] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f21716b8-fbd5-43a7-a468-e41da2cc9b75 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.655161] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c887c097-25e1-4360-81da-1a1301cef338 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.677579] env[62914]: DEBUG nova.compute.provider_tree [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.680079] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 984.680291] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 984.680499] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleting the datastore file [datastore2] 412c706a-31ce-4a33-914d-5762b02b4c63 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.680932] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5276b9b5-1de2-4654-b66f-b161e70cad2b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.686520] env[62914]: DEBUG oslo_vmware.api [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 984.686520] env[62914]: value = "task-1352880" [ 984.686520] env[62914]: _type = "Task" [ 984.686520] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.694881] env[62914]: DEBUG oslo_vmware.api [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352880, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.002664] env[62914]: DEBUG nova.compute.manager [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 985.062730] env[62914]: DEBUG oslo_vmware.api [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352877, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.108335] env[62914]: DEBUG nova.compute.utils [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 985.182355] env[62914]: DEBUG nova.scheduler.client.report [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.196498] env[62914]: DEBUG oslo_vmware.api [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352880, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134172} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.196828] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.197091] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 985.197293] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 985.197472] env[62914]: INFO nova.compute.manager [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Took 1.12 seconds to destroy the instance on the hypervisor. [ 985.197719] env[62914]: DEBUG oslo.service.loopingcall [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.197915] env[62914]: DEBUG nova.compute.manager [-] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 985.198023] env[62914]: DEBUG nova.network.neutron [-] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 985.526629] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.562887] env[62914]: DEBUG oslo_vmware.api [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352877, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.611212] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "2575bc0d-cb47-4dab-a219-f84254bda47d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.689431] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.682s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.692027] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.474s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.693247] env[62914]: INFO nova.compute.claims [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 985.710728] env[62914]: INFO nova.scheduler.client.report [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleted allocations for instance 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509 [ 985.834454] env[62914]: DEBUG nova.compute.manager [req-3959deac-c743-4077-9bf7-7cc2f8d15abe req-2c853641-99c1-4364-abd9-417beebd7d78 service nova] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Received event network-vif-deleted-650bc64b-7970-4ad4-8aa6-0065c010c7d4 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.834663] env[62914]: INFO nova.compute.manager [req-3959deac-c743-4077-9bf7-7cc2f8d15abe req-2c853641-99c1-4364-abd9-417beebd7d78 service nova] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Neutron deleted interface 650bc64b-7970-4ad4-8aa6-0065c010c7d4; detaching it from the instance and deleting it from the info cache [ 985.834970] env[62914]: DEBUG nova.network.neutron [req-3959deac-c743-4077-9bf7-7cc2f8d15abe req-2c853641-99c1-4364-abd9-417beebd7d78 service nova] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.937387] env[62914]: DEBUG nova.network.neutron [-] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.063466] env[62914]: DEBUG oslo_vmware.api [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352877, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.223855] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a4fbc91d-1f7d-4854-8571-1033a7aa9322 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.195s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.337253] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-46d4f659-aa42-48b3-bf55-66614f67acd8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.346558] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5edaa3e-330d-41aa-b934-7fe87bded090 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.372677] env[62914]: DEBUG nova.compute.manager [req-3959deac-c743-4077-9bf7-7cc2f8d15abe req-2c853641-99c1-4364-abd9-417beebd7d78 service nova] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Detach interface failed, port_id=650bc64b-7970-4ad4-8aa6-0065c010c7d4, reason: Instance 412c706a-31ce-4a33-914d-5762b02b4c63 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 986.440097] env[62914]: INFO nova.compute.manager [-] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Took 1.24 seconds to deallocate network for instance. [ 986.564859] env[62914]: DEBUG oslo_vmware.api [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352877, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.671574] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "2575bc0d-cb47-4dab-a219-f84254bda47d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.671860] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "2575bc0d-cb47-4dab-a219-f84254bda47d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.672126] env[62914]: INFO nova.compute.manager [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Attaching volume 5c35f663-6163-4c84-9fc5-1cdcb747f2b3 to /dev/sdb [ 986.706732] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2273ba42-42e4-494e-a6cf-074ceb1c2207 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.714132] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915eb26b-23e9-4163-af5b-c5b9bb98d093 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.729951] env[62914]: DEBUG nova.virt.block_device [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Updating existing volume attachment record: 8fceb884-cac4-4de9-bbe8-a1c489e92c1a {{(pid=62914) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 986.809349] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc9b875-7b1f-4178-8148-fdd83c7b7345 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.816915] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedd0a79-2272-447c-a954-2b76e3e59caf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.845693] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f061dde-b5d0-4363-8baf-af1e08502796 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.853884] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb01075-2f02-404e-8856-84f72407aff0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.865718] env[62914]: DEBUG nova.compute.provider_tree [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.870606] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.870830] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.947399] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.064383] env[62914]: DEBUG oslo_vmware.api [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352877, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.371612] env[62914]: DEBUG nova.scheduler.client.report [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.374809] env[62914]: DEBUG nova.compute.manager [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 987.565209] env[62914]: DEBUG oslo_vmware.api [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352877, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.879008] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.187s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.879599] env[62914]: DEBUG nova.compute.manager [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 987.884891] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.358s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.886201] env[62914]: INFO nova.compute.claims [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 987.906587] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.066211] env[62914]: DEBUG oslo_vmware.api [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352877, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.387049] env[62914]: DEBUG nova.compute.utils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 988.387643] env[62914]: DEBUG nova.compute.manager [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 988.387815] env[62914]: DEBUG nova.network.neutron [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 988.425088] env[62914]: DEBUG nova.policy [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2c21c59e39b94deea6c3701fe971198a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1dcda44f0d6a455bb17966e7c5061190', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 988.566532] env[62914]: DEBUG oslo_vmware.api [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352877, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.680554] env[62914]: DEBUG nova.network.neutron [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Successfully created port: d04462d2-47dc-4db8-838d-0912b80a1db3 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 988.891475] env[62914]: DEBUG nova.compute.manager [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 989.057106] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af7c2f96-7520-4374-a9d4-6b16ff25d20b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.069089] env[62914]: DEBUG oslo_vmware.api [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352877, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.072795] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f797cb8e-400b-45c2-8339-284702df2bc0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.108662] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d22610b-a3fb-485c-aea9-04ad32c1adcd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.121022] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ec86a6-f556-4c06-bf1c-8db04e991754 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.132750] env[62914]: DEBUG nova.compute.provider_tree [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.569149] env[62914]: DEBUG oslo_vmware.api [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352877, 'name': ReconfigVM_Task, 'duration_secs': 5.735279} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.569657] env[62914]: DEBUG oslo_concurrency.lockutils [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.569876] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Reconfigured VM to detach interface {{(pid=62914) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 989.637124] env[62914]: DEBUG nova.scheduler.client.report [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 989.828366] env[62914]: DEBUG nova.compute.manager [req-6f9e8c31-56cd-4d16-ac33-eec47bc544ad req-6c9a25e1-18ba-49cd-99aa-5980665c7698 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Received event network-vif-deleted-3e7456bf-2578-404e-98da-f253c33ac065 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.828366] env[62914]: INFO nova.compute.manager [req-6f9e8c31-56cd-4d16-ac33-eec47bc544ad req-6c9a25e1-18ba-49cd-99aa-5980665c7698 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Neutron deleted interface 3e7456bf-2578-404e-98da-f253c33ac065; detaching it from the instance and deleting it from the info cache [ 989.828706] env[62914]: DEBUG nova.network.neutron [req-6f9e8c31-56cd-4d16-ac33-eec47bc544ad req-6c9a25e1-18ba-49cd-99aa-5980665c7698 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updating instance_info_cache with network_info: [{"id": "615c8b1c-438d-43fb-999b-89172ade3f61", "address": "fa:16:3e:9d:b8:15", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap615c8b1c-43", "ovs_interfaceid": "615c8b1c-438d-43fb-999b-89172ade3f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "285c9a22-ca86-4373-8394-fab932eddd6f", "address": "fa:16:3e:58:af:43", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap285c9a22-ca", "ovs_interfaceid": "285c9a22-ca86-4373-8394-fab932eddd6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.905061] env[62914]: DEBUG nova.compute.manager [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 989.954666] env[62914]: DEBUG nova.virt.hardware [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 989.954995] env[62914]: DEBUG nova.virt.hardware [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 989.955225] env[62914]: DEBUG nova.virt.hardware [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 989.955441] env[62914]: DEBUG nova.virt.hardware [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 989.955600] env[62914]: DEBUG nova.virt.hardware [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 989.955754] env[62914]: DEBUG nova.virt.hardware [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 989.955969] env[62914]: DEBUG nova.virt.hardware [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 989.956151] env[62914]: DEBUG nova.virt.hardware [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 989.956329] env[62914]: DEBUG nova.virt.hardware [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 989.956510] env[62914]: DEBUG nova.virt.hardware [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 989.956725] env[62914]: DEBUG nova.virt.hardware [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 989.957603] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97eff58-4aab-4e26-8211-f4c3231458ba {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.966135] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56ff20d-2bcc-4d4e-9410-8b6c474f4077 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.142492] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.143108] env[62914]: DEBUG nova.compute.manager [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 990.145685] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.198s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.145911] env[62914]: DEBUG nova.objects.instance [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lazy-loading 'resources' on Instance uuid 412c706a-31ce-4a33-914d-5762b02b4c63 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.331631] env[62914]: DEBUG oslo_concurrency.lockutils [req-6f9e8c31-56cd-4d16-ac33-eec47bc544ad req-6c9a25e1-18ba-49cd-99aa-5980665c7698 service nova] Acquiring lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.331822] env[62914]: DEBUG oslo_concurrency.lockutils [req-6f9e8c31-56cd-4d16-ac33-eec47bc544ad req-6c9a25e1-18ba-49cd-99aa-5980665c7698 service nova] Acquired lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.332724] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fa2126-6536-4f96-b854-ef911bf26915 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.353631] env[62914]: DEBUG oslo_concurrency.lockutils [req-6f9e8c31-56cd-4d16-ac33-eec47bc544ad req-6c9a25e1-18ba-49cd-99aa-5980665c7698 service nova] Releasing lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.353952] env[62914]: WARNING nova.compute.manager [req-6f9e8c31-56cd-4d16-ac33-eec47bc544ad req-6c9a25e1-18ba-49cd-99aa-5980665c7698 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Detach interface failed, port_id=3e7456bf-2578-404e-98da-f253c33ac065, reason: No device with interface-id 3e7456bf-2578-404e-98da-f253c33ac065 exists on VM: nova.exception.NotFound: No device with interface-id 3e7456bf-2578-404e-98da-f253c33ac065 exists on VM [ 990.358230] env[62914]: DEBUG nova.compute.manager [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Received event network-vif-deleted-285c9a22-ca86-4373-8394-fab932eddd6f {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.358415] env[62914]: INFO nova.compute.manager [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Neutron deleted interface 285c9a22-ca86-4373-8394-fab932eddd6f; detaching it from the instance and deleting it from the info cache [ 990.358661] env[62914]: DEBUG nova.network.neutron [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updating instance_info_cache with network_info: [{"id": "615c8b1c-438d-43fb-999b-89172ade3f61", "address": "fa:16:3e:9d:b8:15", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap615c8b1c-43", "ovs_interfaceid": "615c8b1c-438d-43fb-999b-89172ade3f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.608072] env[62914]: DEBUG nova.network.neutron [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Successfully updated port: d04462d2-47dc-4db8-838d-0912b80a1db3 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 990.649811] env[62914]: DEBUG nova.compute.utils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 990.657028] env[62914]: DEBUG nova.compute.manager [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 990.657028] env[62914]: DEBUG nova.network.neutron [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 990.715774] env[62914]: DEBUG nova.policy [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41c19b40819c49d8a9dacdb74ad5899b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f5fa79f0260e461c952301f904101f79', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 990.767499] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-051882f4-0564-463a-8842-d41e8bd3ca1f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.775456] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1774b9f-4f39-4905-93d4-e72ae03537f3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.805771] env[62914]: DEBUG oslo_concurrency.lockutils [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.805958] env[62914]: DEBUG oslo_concurrency.lockutils [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.806181] env[62914]: DEBUG nova.network.neutron [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.808220] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da417628-e3ce-490f-9c68-f38dde8e0ead {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.815591] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c30642-5bf5-48bb-8f78-1277a193c72e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.829576] env[62914]: DEBUG nova.compute.provider_tree [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.861061] env[62914]: DEBUG oslo_concurrency.lockutils [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Acquiring lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.861061] env[62914]: DEBUG oslo_concurrency.lockutils [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Acquired lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.861491] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-501abd74-1f22-4d92-8c72-9d34f8c45091 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.878615] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db337280-15fb-423e-946d-024db7236cea {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.902815] env[62914]: DEBUG nova.virt.vmwareapi.vmops [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Reconfiguring VM to detach interface {{(pid=62914) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 990.903080] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1776e741-9330-4243-aa5a-1d347b5bce36 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.915787] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "1e527707-758c-4006-af83-9c739b9645ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.921029] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Waiting for the task: (returnval){ [ 990.921029] env[62914]: value = "task-1352883" [ 990.921029] env[62914]: _type = "Task" [ 990.921029] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.930504] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.960125] env[62914]: DEBUG nova.network.neutron [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Successfully created port: fadaceea-8eb6-4ab7-9c90-8095a38ec8cf {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 991.083363] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "c7cf71d7-b823-471f-949f-9ef2b539d972" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.083363] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.083363] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "c7cf71d7-b823-471f-949f-9ef2b539d972-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.083363] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.083363] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.085658] env[62914]: INFO nova.compute.manager [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Terminating instance [ 991.087663] env[62914]: DEBUG nova.compute.manager [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 991.087873] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 991.088699] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05884262-6eec-4d35-8d80-6427c8501ff3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.096606] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 991.096852] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9af32f84-7966-4f59-bd71-a71cb32f06aa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.103346] env[62914]: DEBUG oslo_vmware.api [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 991.103346] env[62914]: value = "task-1352884" [ 991.103346] env[62914]: _type = "Task" [ 991.103346] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.111202] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "refresh_cache-84a7822d-d46d-4147-ad4c-b9402aa072f4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.111336] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "refresh_cache-84a7822d-d46d-4147-ad4c-b9402aa072f4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.111482] env[62914]: DEBUG nova.network.neutron [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 991.113444] env[62914]: DEBUG oslo_vmware.api [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352884, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.155145] env[62914]: DEBUG nova.compute.manager [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 991.274416] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Volume attach. Driver type: vmdk {{(pid=62914) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 991.274714] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288258', 'volume_id': '5c35f663-6163-4c84-9fc5-1cdcb747f2b3', 'name': 'volume-5c35f663-6163-4c84-9fc5-1cdcb747f2b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2575bc0d-cb47-4dab-a219-f84254bda47d', 'attached_at': '', 'detached_at': '', 'volume_id': '5c35f663-6163-4c84-9fc5-1cdcb747f2b3', 'serial': '5c35f663-6163-4c84-9fc5-1cdcb747f2b3'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 991.275652] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8cf5c7-db06-43eb-ab8c-6a6e0179092e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.291900] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-710e40f3-34fc-4fd2-bc78-98039427edd0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.318128] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] volume-5c35f663-6163-4c84-9fc5-1cdcb747f2b3/volume-5c35f663-6163-4c84-9fc5-1cdcb747f2b3.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.318415] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8b53f1e-e3b4-438f-9a7f-910b71b2f530 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.332555] env[62914]: DEBUG nova.scheduler.client.report [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 991.341491] env[62914]: DEBUG oslo_vmware.api [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 991.341491] env[62914]: value = "task-1352885" [ 991.341491] env[62914]: _type = "Task" [ 991.341491] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.349558] env[62914]: DEBUG oslo_vmware.api [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352885, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.431389] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.614432] env[62914]: DEBUG oslo_vmware.api [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352884, 'name': PowerOffVM_Task, 'duration_secs': 0.190285} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.614795] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 991.615055] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 991.617093] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-48790d53-31c0-4b42-aba1-f9dd584b9aa2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.677333] env[62914]: DEBUG nova.network.neutron [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 991.694622] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 991.694861] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 991.695090] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Deleting the datastore file [datastore1] c7cf71d7-b823-471f-949f-9ef2b539d972 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.695368] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ccf50c0-66a8-4441-9ab0-91f9970413e3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.705434] env[62914]: DEBUG oslo_vmware.api [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for the task: (returnval){ [ 991.705434] env[62914]: value = "task-1352887" [ 991.705434] env[62914]: _type = "Task" [ 991.705434] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.715930] env[62914]: DEBUG oslo_vmware.api [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352887, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.837609] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.692s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.841013] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.935s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.842769] env[62914]: INFO nova.compute.claims [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 991.858465] env[62914]: DEBUG oslo_vmware.api [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352885, 'name': ReconfigVM_Task, 'duration_secs': 0.325863} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.858750] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Reconfigured VM instance instance-0000005b to attach disk [datastore2] volume-5c35f663-6163-4c84-9fc5-1cdcb747f2b3/volume-5c35f663-6163-4c84-9fc5-1cdcb747f2b3.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 991.863909] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6c9b8b9-b97b-479a-9708-740a220e1c6f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.874294] env[62914]: INFO nova.scheduler.client.report [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleted allocations for instance 412c706a-31ce-4a33-914d-5762b02b4c63 [ 991.882722] env[62914]: DEBUG oslo_vmware.api [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 991.882722] env[62914]: value = "task-1352888" [ 991.882722] env[62914]: _type = "Task" [ 991.882722] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.892800] env[62914]: DEBUG oslo_vmware.api [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352888, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.932632] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.959603] env[62914]: DEBUG nova.network.neutron [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updating instance_info_cache with network_info: [{"id": "615c8b1c-438d-43fb-999b-89172ade3f61", "address": "fa:16:3e:9d:b8:15", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap615c8b1c-43", "ovs_interfaceid": "615c8b1c-438d-43fb-999b-89172ade3f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.976218] env[62914]: DEBUG nova.network.neutron [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Updating instance_info_cache with network_info: [{"id": "d04462d2-47dc-4db8-838d-0912b80a1db3", "address": "fa:16:3e:62:95:46", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04462d2-47", "ovs_interfaceid": "d04462d2-47dc-4db8-838d-0912b80a1db3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.165174] env[62914]: DEBUG nova.compute.manager [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 992.191483] env[62914]: DEBUG nova.virt.hardware [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 992.191706] env[62914]: DEBUG nova.virt.hardware [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 992.191870] env[62914]: DEBUG nova.virt.hardware [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 992.192067] env[62914]: DEBUG nova.virt.hardware [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 992.192225] env[62914]: DEBUG nova.virt.hardware [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 992.192373] env[62914]: DEBUG nova.virt.hardware [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 992.192577] env[62914]: DEBUG nova.virt.hardware [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 992.192839] env[62914]: DEBUG nova.virt.hardware [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 992.193045] env[62914]: DEBUG nova.virt.hardware [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 992.193221] env[62914]: DEBUG nova.virt.hardware [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 992.193397] env[62914]: DEBUG nova.virt.hardware [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 992.194340] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39552f78-78ca-4e11-8cc2-8d1b22ce7714 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.203163] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa6fe63-ab2b-4fea-84f1-2b9491daa39b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.214814] env[62914]: DEBUG oslo_vmware.api [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Task: {'id': task-1352887, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143707} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.221814] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.222019] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 992.222200] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 992.222375] env[62914]: INFO nova.compute.manager [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Took 1.13 seconds to destroy the instance on the hypervisor. [ 992.222611] env[62914]: DEBUG oslo.service.loopingcall [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.223041] env[62914]: DEBUG nova.compute.manager [-] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 992.223148] env[62914]: DEBUG nova.network.neutron [-] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 992.383974] env[62914]: DEBUG oslo_concurrency.lockutils [None req-d7be4768-beda-418e-b3c9-39581c2d68bd tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "412c706a-31ce-4a33-914d-5762b02b4c63" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.314s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.395635] env[62914]: DEBUG oslo_vmware.api [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352888, 'name': ReconfigVM_Task, 'duration_secs': 0.128254} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.397535] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288258', 'volume_id': '5c35f663-6163-4c84-9fc5-1cdcb747f2b3', 'name': 'volume-5c35f663-6163-4c84-9fc5-1cdcb747f2b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2575bc0d-cb47-4dab-a219-f84254bda47d', 'attached_at': '', 'detached_at': '', 'volume_id': '5c35f663-6163-4c84-9fc5-1cdcb747f2b3', 'serial': '5c35f663-6163-4c84-9fc5-1cdcb747f2b3'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 992.401191] env[62914]: DEBUG nova.compute.manager [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Received event network-vif-plugged-d04462d2-47dc-4db8-838d-0912b80a1db3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.401191] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] Acquiring lock "84a7822d-d46d-4147-ad4c-b9402aa072f4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.401191] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] Lock "84a7822d-d46d-4147-ad4c-b9402aa072f4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.401393] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] Lock "84a7822d-d46d-4147-ad4c-b9402aa072f4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.402025] env[62914]: DEBUG nova.compute.manager [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] No waiting events found dispatching network-vif-plugged-d04462d2-47dc-4db8-838d-0912b80a1db3 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 992.402025] env[62914]: WARNING nova.compute.manager [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Received unexpected event network-vif-plugged-d04462d2-47dc-4db8-838d-0912b80a1db3 for instance with vm_state building and task_state spawning. [ 992.402025] env[62914]: DEBUG nova.compute.manager [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Received event network-changed-d04462d2-47dc-4db8-838d-0912b80a1db3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.402025] env[62914]: DEBUG nova.compute.manager [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Refreshing instance network info cache due to event network-changed-d04462d2-47dc-4db8-838d-0912b80a1db3. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 992.402257] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] Acquiring lock "refresh_cache-84a7822d-d46d-4147-ad4c-b9402aa072f4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.434419] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.440857] env[62914]: DEBUG nova.network.neutron [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Successfully updated port: fadaceea-8eb6-4ab7-9c90-8095a38ec8cf {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 992.462737] env[62914]: DEBUG oslo_concurrency.lockutils [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "refresh_cache-1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.478605] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "refresh_cache-84a7822d-d46d-4147-ad4c-b9402aa072f4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.478912] env[62914]: DEBUG nova.compute.manager [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Instance network_info: |[{"id": "d04462d2-47dc-4db8-838d-0912b80a1db3", "address": "fa:16:3e:62:95:46", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04462d2-47", "ovs_interfaceid": "d04462d2-47dc-4db8-838d-0912b80a1db3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 992.479498] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] Acquired lock "refresh_cache-84a7822d-d46d-4147-ad4c-b9402aa072f4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.480777] env[62914]: DEBUG nova.network.neutron [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Refreshing network info cache for port d04462d2-47dc-4db8-838d-0912b80a1db3 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 992.480868] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:95:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a8c8175-1197-4f12-baac-ef6aba95f585', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd04462d2-47dc-4db8-838d-0912b80a1db3', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 992.488634] env[62914]: DEBUG oslo.service.loopingcall [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.491226] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 992.491677] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-60a0a886-3f70-40dc-803a-48f5883b979e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.513071] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 992.513071] env[62914]: value = "task-1352889" [ 992.513071] env[62914]: _type = "Task" [ 992.513071] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.520690] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352889, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.770601] env[62914]: DEBUG nova.compute.manager [req-dd720325-8815-4cd5-aae9-5da516c68aea req-0173baa6-248d-4b68-9647-06ce431e7eda service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Received event network-vif-deleted-57172af7-5ed6-46d9-9d7b-3ef100c530f7 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.771513] env[62914]: INFO nova.compute.manager [req-dd720325-8815-4cd5-aae9-5da516c68aea req-0173baa6-248d-4b68-9647-06ce431e7eda service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Neutron deleted interface 57172af7-5ed6-46d9-9d7b-3ef100c530f7; detaching it from the instance and deleting it from the info cache [ 992.773400] env[62914]: DEBUG nova.network.neutron [req-dd720325-8815-4cd5-aae9-5da516c68aea req-0173baa6-248d-4b68-9647-06ce431e7eda service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.777706] env[62914]: DEBUG nova.network.neutron [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Updated VIF entry in instance network info cache for port d04462d2-47dc-4db8-838d-0912b80a1db3. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.778077] env[62914]: DEBUG nova.network.neutron [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Updating instance_info_cache with network_info: [{"id": "d04462d2-47dc-4db8-838d-0912b80a1db3", "address": "fa:16:3e:62:95:46", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04462d2-47", "ovs_interfaceid": "d04462d2-47dc-4db8-838d-0912b80a1db3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.933948] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.947791] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "refresh_cache-054f37ad-4161-4c70-aa1b-b2e69cd1e9f4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.947916] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "refresh_cache-054f37ad-4161-4c70-aa1b-b2e69cd1e9f4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.948077] env[62914]: DEBUG nova.network.neutron [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 992.966089] env[62914]: DEBUG oslo_concurrency.lockutils [None req-38917157-e757-47c0-81a2-35523cf17658 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-1e527707-758c-4006-af83-9c739b9645ed-3e7456bf-2578-404e-98da-f253c33ac065" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.994s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.968600] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb089ef8-ccf4-4461-b582-0b452482a393 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.976595] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79938f25-3edd-473c-86ae-c90390cadb01 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.008965] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e83ff1-44fc-4c6d-9e4b-16fee0ef0e99 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.018431] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e17308-c1c6-4d78-b488-efcea240d1d1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.028217] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352889, 'name': CreateVM_Task, 'duration_secs': 0.374695} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.035493] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 993.035943] env[62914]: DEBUG nova.compute.provider_tree [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.038029] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.038282] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.038600] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 993.039232] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13d3f54e-ef5d-4ce7-ae7d-3ef2b138f07b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.043595] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 993.043595] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52b60847-30ef-13af-008c-cf7e6281d3a9" [ 993.043595] env[62914]: _type = "Task" [ 993.043595] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.051879] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52b60847-30ef-13af-008c-cf7e6281d3a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.177964] env[62914]: DEBUG nova.network.neutron [-] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.276023] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-11cac69b-9a21-431a-be8f-704a378e03d0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.280827] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] Releasing lock "refresh_cache-84a7822d-d46d-4147-ad4c-b9402aa072f4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.281087] env[62914]: DEBUG nova.compute.manager [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Received event network-vif-plugged-fadaceea-8eb6-4ab7-9c90-8095a38ec8cf {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.281293] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] Acquiring lock "054f37ad-4161-4c70-aa1b-b2e69cd1e9f4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.281493] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] Lock "054f37ad-4161-4c70-aa1b-b2e69cd1e9f4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.281647] env[62914]: DEBUG oslo_concurrency.lockutils [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] Lock "054f37ad-4161-4c70-aa1b-b2e69cd1e9f4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.281818] env[62914]: DEBUG nova.compute.manager [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] No waiting events found dispatching network-vif-plugged-fadaceea-8eb6-4ab7-9c90-8095a38ec8cf {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 993.282028] env[62914]: WARNING nova.compute.manager [req-8b42f42b-f817-48ef-b482-3e6d594e0bc6 req-16405e6b-0701-4b26-9fa8-8dd47197883f service nova] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Received unexpected event network-vif-plugged-fadaceea-8eb6-4ab7-9c90-8095a38ec8cf for instance with vm_state building and task_state spawning. [ 993.285871] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b6d236-542f-4a82-b687-5c8696180f99 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.310269] env[62914]: DEBUG nova.compute.manager [req-dd720325-8815-4cd5-aae9-5da516c68aea req-0173baa6-248d-4b68-9647-06ce431e7eda service nova] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Detach interface failed, port_id=57172af7-5ed6-46d9-9d7b-3ef100c530f7, reason: Instance c7cf71d7-b823-471f-949f-9ef2b539d972 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 993.433400] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.448902] env[62914]: DEBUG nova.objects.instance [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lazy-loading 'flavor' on Instance uuid 2575bc0d-cb47-4dab-a219-f84254bda47d {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.478935] env[62914]: DEBUG nova.network.neutron [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 993.538695] env[62914]: DEBUG nova.scheduler.client.report [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.554734] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52b60847-30ef-13af-008c-cf7e6281d3a9, 'name': SearchDatastore_Task, 'duration_secs': 0.008804} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.555739] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.556025] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 993.556285] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.556440] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.556619] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 993.557179] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea5c5d94-41d8-4ab2-9db5-ce9c9fc916c8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.565341] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 993.565341] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 993.565966] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-439a4f20-0cc7-4e50-9fb7-7a05658685bb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.570960] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 993.570960] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52404f68-c51c-fd35-50af-27832e385dfb" [ 993.570960] env[62914]: _type = "Task" [ 993.570960] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.578102] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52404f68-c51c-fd35-50af-27832e385dfb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.616254] env[62914]: DEBUG nova.network.neutron [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Updating instance_info_cache with network_info: [{"id": "fadaceea-8eb6-4ab7-9c90-8095a38ec8cf", "address": "fa:16:3e:0f:1b:f4", "network": {"id": "c9c30d64-cafe-48a5-b868-b20d90d8920d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-458611021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5fa79f0260e461c952301f904101f79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfadaceea-8e", "ovs_interfaceid": "fadaceea-8eb6-4ab7-9c90-8095a38ec8cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.681538] env[62914]: INFO nova.compute.manager [-] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Took 1.46 seconds to deallocate network for instance. [ 993.934657] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.955694] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6aa6285b-2250-4c22-b759-1a8eb3f3c730 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "2575bc0d-cb47-4dab-a219-f84254bda47d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.284s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.043759] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.203s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.044474] env[62914]: DEBUG nova.compute.manager [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 994.082557] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52404f68-c51c-fd35-50af-27832e385dfb, 'name': SearchDatastore_Task, 'duration_secs': 0.007552} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.083395] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-446bc485-d56f-4e00-8380-0c4c5f1cafe9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.089682] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 994.089682] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]526cc543-189e-0d44-6f29-1207ff21ff60" [ 994.089682] env[62914]: _type = "Task" [ 994.089682] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.097722] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]526cc543-189e-0d44-6f29-1207ff21ff60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.118604] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "refresh_cache-054f37ad-4161-4c70-aa1b-b2e69cd1e9f4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.118899] env[62914]: DEBUG nova.compute.manager [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Instance network_info: |[{"id": "fadaceea-8eb6-4ab7-9c90-8095a38ec8cf", "address": "fa:16:3e:0f:1b:f4", "network": {"id": "c9c30d64-cafe-48a5-b868-b20d90d8920d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-458611021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5fa79f0260e461c952301f904101f79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfadaceea-8e", "ovs_interfaceid": "fadaceea-8eb6-4ab7-9c90-8095a38ec8cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 994.119309] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0f:1b:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fadaceea-8eb6-4ab7-9c90-8095a38ec8cf', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 994.126679] env[62914]: DEBUG oslo.service.loopingcall [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.126995] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "2575bc0d-cb47-4dab-a219-f84254bda47d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.127236] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "2575bc0d-cb47-4dab-a219-f84254bda47d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.128452] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 994.128811] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fe00ab0a-1b4e-406e-ab1a-f13d373f96e2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.151077] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 994.151077] env[62914]: value = "task-1352890" [ 994.151077] env[62914]: _type = "Task" [ 994.151077] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.161151] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352890, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.187776] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.188090] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.188317] env[62914]: DEBUG nova.objects.instance [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lazy-loading 'resources' on Instance uuid c7cf71d7-b823-471f-949f-9ef2b539d972 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 994.419232] env[62914]: DEBUG nova.compute.manager [req-1e957cba-26d1-4727-878f-cee7d7a4873c req-7336ff89-08bc-4e27-9297-435467a29f0d service nova] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Received event network-changed-fadaceea-8eb6-4ab7-9c90-8095a38ec8cf {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.419467] env[62914]: DEBUG nova.compute.manager [req-1e957cba-26d1-4727-878f-cee7d7a4873c req-7336ff89-08bc-4e27-9297-435467a29f0d service nova] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Refreshing instance network info cache due to event network-changed-fadaceea-8eb6-4ab7-9c90-8095a38ec8cf. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 994.419633] env[62914]: DEBUG oslo_concurrency.lockutils [req-1e957cba-26d1-4727-878f-cee7d7a4873c req-7336ff89-08bc-4e27-9297-435467a29f0d service nova] Acquiring lock "refresh_cache-054f37ad-4161-4c70-aa1b-b2e69cd1e9f4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.419783] env[62914]: DEBUG oslo_concurrency.lockutils [req-1e957cba-26d1-4727-878f-cee7d7a4873c req-7336ff89-08bc-4e27-9297-435467a29f0d service nova] Acquired lock "refresh_cache-054f37ad-4161-4c70-aa1b-b2e69cd1e9f4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.419974] env[62914]: DEBUG nova.network.neutron [req-1e957cba-26d1-4727-878f-cee7d7a4873c req-7336ff89-08bc-4e27-9297-435467a29f0d service nova] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Refreshing network info cache for port fadaceea-8eb6-4ab7-9c90-8095a38ec8cf {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 994.436450] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.549554] env[62914]: DEBUG nova.compute.utils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 994.551149] env[62914]: DEBUG nova.compute.manager [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 994.551327] env[62914]: DEBUG nova.network.neutron [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 994.595026] env[62914]: DEBUG nova.policy [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '795b94761b144a1e99ed79838d9687ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9278818cebd842669b802a4e43e91774', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 994.602583] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]526cc543-189e-0d44-6f29-1207ff21ff60, 'name': SearchDatastore_Task, 'duration_secs': 0.009035} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.602891] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.603163] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 84a7822d-d46d-4147-ad4c-b9402aa072f4/84a7822d-d46d-4147-ad4c-b9402aa072f4.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 994.603420] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1c7d3113-8ac1-4c1c-b595-87c95ce68b0c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.609649] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 994.609649] env[62914]: value = "task-1352891" [ 994.609649] env[62914]: _type = "Task" [ 994.609649] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.617301] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352891, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.629441] env[62914]: INFO nova.compute.manager [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Detaching volume 5c35f663-6163-4c84-9fc5-1cdcb747f2b3 [ 994.661378] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352890, 'name': CreateVM_Task, 'duration_secs': 0.262778} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.661551] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 994.662215] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.662387] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.662784] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 994.662968] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c086283b-3240-4ca5-a8a3-a12d6a730540 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.665727] env[62914]: INFO nova.virt.block_device [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Attempting to driver detach volume 5c35f663-6163-4c84-9fc5-1cdcb747f2b3 from mountpoint /dev/sdb [ 994.665925] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Volume detach. Driver type: vmdk {{(pid=62914) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 994.666139] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288258', 'volume_id': '5c35f663-6163-4c84-9fc5-1cdcb747f2b3', 'name': 'volume-5c35f663-6163-4c84-9fc5-1cdcb747f2b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2575bc0d-cb47-4dab-a219-f84254bda47d', 'attached_at': '', 'detached_at': '', 'volume_id': '5c35f663-6163-4c84-9fc5-1cdcb747f2b3', 'serial': '5c35f663-6163-4c84-9fc5-1cdcb747f2b3'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 994.666904] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6a54e4-4291-4bb6-a637-c851f2eaaabf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.670324] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 994.670324] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52275a22-906c-f239-1283-4bb19548e4c0" [ 994.670324] env[62914]: _type = "Task" [ 994.670324] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.698028] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c51d8534-ba88-48de-be75-020f72997fbc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.700852] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52275a22-906c-f239-1283-4bb19548e4c0, 'name': SearchDatastore_Task, 'duration_secs': 0.009495} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.701151] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.701396] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 994.701653] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.701805] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.701987] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 994.702539] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b35820df-f096-476a-ad83-d8e51f98b7ea {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.706474] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ca79db-4e0c-47f3-9a56-53fd8ad86fcd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.710268] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 994.710442] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 994.728148] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbcb6d9c-af21-411d-a72e-f9c9c49e3b26 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.730989] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d705f83-f35c-4cef-bad2-7a75dfcdc342 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.735879] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 994.735879] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c194d2-53b2-b56f-353d-6a065bc12a29" [ 994.735879] env[62914]: _type = "Task" [ 994.735879] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.746086] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] The volume has not been displaced from its original location: [datastore2] volume-5c35f663-6163-4c84-9fc5-1cdcb747f2b3/volume-5c35f663-6163-4c84-9fc5-1cdcb747f2b3.vmdk. No consolidation needed. {{(pid=62914) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 994.751930] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Reconfiguring VM instance instance-0000005b to detach disk 2001 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 994.757371] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0e8e4d5-c960-4d5e-a14f-e799a960da0c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.776271] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52c194d2-53b2-b56f-353d-6a065bc12a29, 'name': SearchDatastore_Task, 'duration_secs': 0.008049} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.778178] env[62914]: DEBUG oslo_vmware.api [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 994.778178] env[62914]: value = "task-1352892" [ 994.778178] env[62914]: _type = "Task" [ 994.778178] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.778406] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b254bc5f-db12-42ee-aaeb-d8954bcd35fc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.790184] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 994.790184] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5214c28a-60b1-5bcf-270d-09e4d9bd3526" [ 994.790184] env[62914]: _type = "Task" [ 994.790184] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.793138] env[62914]: DEBUG oslo_vmware.api [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352892, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.803935] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5214c28a-60b1-5bcf-270d-09e4d9bd3526, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.851368] env[62914]: DEBUG nova.network.neutron [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Successfully created port: 9c9a03a9-3a96-4060-97c8-3fdf53f21417 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 994.856876] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9fe5364-03d6-4cfa-96a9-9a5bbf668908 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.865833] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de0251d-41e6-4725-8af5-463e4d36f8b8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.903315] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb79d25-f531-405b-a8c2-492a2e1670db {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.913204] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e5a85e-9505-41b2-80f5-546291754ad1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.931828] env[62914]: DEBUG nova.compute.provider_tree [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.943667] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.054560] env[62914]: DEBUG nova.compute.manager [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 995.119751] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352891, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477579} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.120052] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 84a7822d-d46d-4147-ad4c-b9402aa072f4/84a7822d-d46d-4147-ad4c-b9402aa072f4.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 995.120273] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 995.120536] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd40efb7-26ae-4d0d-923a-bd01536b110a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.127332] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 995.127332] env[62914]: value = "task-1352893" [ 995.127332] env[62914]: _type = "Task" [ 995.127332] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.138162] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352893, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.247730] env[62914]: DEBUG nova.network.neutron [req-1e957cba-26d1-4727-878f-cee7d7a4873c req-7336ff89-08bc-4e27-9297-435467a29f0d service nova] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Updated VIF entry in instance network info cache for port fadaceea-8eb6-4ab7-9c90-8095a38ec8cf. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 995.248128] env[62914]: DEBUG nova.network.neutron [req-1e957cba-26d1-4727-878f-cee7d7a4873c req-7336ff89-08bc-4e27-9297-435467a29f0d service nova] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Updating instance_info_cache with network_info: [{"id": "fadaceea-8eb6-4ab7-9c90-8095a38ec8cf", "address": "fa:16:3e:0f:1b:f4", "network": {"id": "c9c30d64-cafe-48a5-b868-b20d90d8920d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-458611021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5fa79f0260e461c952301f904101f79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfadaceea-8e", "ovs_interfaceid": "fadaceea-8eb6-4ab7-9c90-8095a38ec8cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.290689] env[62914]: DEBUG oslo_vmware.api [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352892, 'name': ReconfigVM_Task, 'duration_secs': 0.278953} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.290964] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Reconfigured VM instance instance-0000005b to detach disk 2001 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 995.295607] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ca1b24d-7e88-4b14-81a1-bdb14f8d78f8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.313886] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5214c28a-60b1-5bcf-270d-09e4d9bd3526, 'name': SearchDatastore_Task, 'duration_secs': 0.055718} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.315019] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.315277] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4/054f37ad-4161-4c70-aa1b-b2e69cd1e9f4.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 995.315581] env[62914]: DEBUG oslo_vmware.api [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 995.315581] env[62914]: value = "task-1352894" [ 995.315581] env[62914]: _type = "Task" [ 995.315581] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.315785] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f14192a-aa7c-4a09-bf8e-1feccc3c0850 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.325711] env[62914]: DEBUG oslo_vmware.api [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352894, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.326818] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 995.326818] env[62914]: value = "task-1352895" [ 995.326818] env[62914]: _type = "Task" [ 995.326818] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.333856] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352895, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.438154] env[62914]: DEBUG nova.scheduler.client.report [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 995.446667] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.638905] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352893, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.751015] env[62914]: DEBUG oslo_concurrency.lockutils [req-1e957cba-26d1-4727-878f-cee7d7a4873c req-7336ff89-08bc-4e27-9297-435467a29f0d service nova] Releasing lock "refresh_cache-054f37ad-4161-4c70-aa1b-b2e69cd1e9f4" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.827549] env[62914]: DEBUG oslo_vmware.api [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352894, 'name': ReconfigVM_Task, 'duration_secs': 0.140545} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.830907] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288258', 'volume_id': '5c35f663-6163-4c84-9fc5-1cdcb747f2b3', 'name': 'volume-5c35f663-6163-4c84-9fc5-1cdcb747f2b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2575bc0d-cb47-4dab-a219-f84254bda47d', 'attached_at': '', 'detached_at': '', 'volume_id': '5c35f663-6163-4c84-9fc5-1cdcb747f2b3', 'serial': '5c35f663-6163-4c84-9fc5-1cdcb747f2b3'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 995.837801] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352895, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.941974] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.942928] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.755s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.962423] env[62914]: INFO nova.scheduler.client.report [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Deleted allocations for instance c7cf71d7-b823-471f-949f-9ef2b539d972 [ 996.068181] env[62914]: DEBUG nova.compute.manager [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 996.095704] env[62914]: DEBUG nova.virt.hardware [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 996.095841] env[62914]: DEBUG nova.virt.hardware [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 996.095976] env[62914]: DEBUG nova.virt.hardware [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 996.096191] env[62914]: DEBUG nova.virt.hardware [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 996.096345] env[62914]: DEBUG nova.virt.hardware [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 996.096581] env[62914]: DEBUG nova.virt.hardware [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 996.096872] env[62914]: DEBUG nova.virt.hardware [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 996.097067] env[62914]: DEBUG nova.virt.hardware [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 996.097260] env[62914]: DEBUG nova.virt.hardware [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 996.097430] env[62914]: DEBUG nova.virt.hardware [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 996.097606] env[62914]: DEBUG nova.virt.hardware [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 996.098612] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c029bc3-4beb-40b7-98a6-c31e05f6606c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.106585] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4346410d-a7f0-4ffa-8ae5-76476cffc62f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.137388] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352893, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.339540] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352895, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.377770] env[62914]: DEBUG nova.objects.instance [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lazy-loading 'flavor' on Instance uuid 2575bc0d-cb47-4dab-a219-f84254bda47d {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.447047] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.470673] env[62914]: DEBUG oslo_concurrency.lockutils [None req-dabd3778-1117-4977-8a64-f9a084dd7c10 tempest-AttachVolumeShelveTestJSON-2108206369 tempest-AttachVolumeShelveTestJSON-2108206369-project-member] Lock "c7cf71d7-b823-471f-949f-9ef2b539d972" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.388s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.483334] env[62914]: DEBUG nova.compute.manager [req-5bdd3e89-afd3-49e5-8cbc-66d23bd91c49 req-9013e7e4-427e-41e2-9eee-3c74855d7246 service nova] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Received event network-vif-plugged-9c9a03a9-3a96-4060-97c8-3fdf53f21417 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.483548] env[62914]: DEBUG oslo_concurrency.lockutils [req-5bdd3e89-afd3-49e5-8cbc-66d23bd91c49 req-9013e7e4-427e-41e2-9eee-3c74855d7246 service nova] Acquiring lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.484535] env[62914]: DEBUG oslo_concurrency.lockutils [req-5bdd3e89-afd3-49e5-8cbc-66d23bd91c49 req-9013e7e4-427e-41e2-9eee-3c74855d7246 service nova] Lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.484870] env[62914]: DEBUG oslo_concurrency.lockutils [req-5bdd3e89-afd3-49e5-8cbc-66d23bd91c49 req-9013e7e4-427e-41e2-9eee-3c74855d7246 service nova] Lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.485225] env[62914]: DEBUG nova.compute.manager [req-5bdd3e89-afd3-49e5-8cbc-66d23bd91c49 req-9013e7e4-427e-41e2-9eee-3c74855d7246 service nova] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] No waiting events found dispatching network-vif-plugged-9c9a03a9-3a96-4060-97c8-3fdf53f21417 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 996.485517] env[62914]: WARNING nova.compute.manager [req-5bdd3e89-afd3-49e5-8cbc-66d23bd91c49 req-9013e7e4-427e-41e2-9eee-3c74855d7246 service nova] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Received unexpected event network-vif-plugged-9c9a03a9-3a96-4060-97c8-3fdf53f21417 for instance with vm_state building and task_state spawning. [ 996.574289] env[62914]: DEBUG nova.network.neutron [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Successfully updated port: 9c9a03a9-3a96-4060-97c8-3fdf53f21417 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 996.638659] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352893, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.020342} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.638846] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 996.640050] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7aa1840-1bea-4b1e-b206-6506af0b2131 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.666313] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 84a7822d-d46d-4147-ad4c-b9402aa072f4/84a7822d-d46d-4147-ad4c-b9402aa072f4.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 996.666638] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c47c65d0-eab0-4b6c-8f17-d69495c7fa0a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.685785] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 996.685785] env[62914]: value = "task-1352896" [ 996.685785] env[62914]: _type = "Task" [ 996.685785] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.693392] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352896, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.837609] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352895, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.214978} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.837609] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4/054f37ad-4161-4c70-aa1b-b2e69cd1e9f4.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 996.838077] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 996.838077] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f9572e3a-25e1-4d61-ab26-e8c5ea7c028b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.844522] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 996.844522] env[62914]: value = "task-1352897" [ 996.844522] env[62914]: _type = "Task" [ 996.844522] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.852822] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352897, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.944251] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.080790] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "refresh_cache-961280ed-0160-4c4e-bca4-c5f2091e9bb0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.080971] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "refresh_cache-961280ed-0160-4c4e-bca4-c5f2091e9bb0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.081144] env[62914]: DEBUG nova.network.neutron [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 997.196047] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352896, 'name': ReconfigVM_Task, 'duration_secs': 0.28161} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.196331] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 84a7822d-d46d-4147-ad4c-b9402aa072f4/84a7822d-d46d-4147-ad4c-b9402aa072f4.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.196975] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ca0139ec-9aa9-4394-85a5-4ba77d3a9bb8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.204538] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 997.204538] env[62914]: value = "task-1352898" [ 997.204538] env[62914]: _type = "Task" [ 997.204538] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.212710] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352898, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.354966] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352897, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063354} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.355401] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 997.356087] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fae358e-6ede-4d9e-843f-52a53e9bf70f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.377676] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4/054f37ad-4161-4c70-aa1b-b2e69cd1e9f4.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.377955] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff12ddd0-726c-4bdd-99e3-2a2f1c3b230a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.392475] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3272709f-e2e6-4e03-91ee-7cc644e0f05e tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "2575bc0d-cb47-4dab-a219-f84254bda47d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.265s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.399622] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 997.399622] env[62914]: value = "task-1352899" [ 997.399622] env[62914]: _type = "Task" [ 997.399622] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.409719] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352899, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.443936] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.612330] env[62914]: DEBUG nova.network.neutron [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 997.716403] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352898, 'name': Rename_Task, 'duration_secs': 0.152024} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.716729] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 997.717071] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-21a1bc72-7796-44be-ab49-a96f9c3b2c1c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.729606] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 997.729606] env[62914]: value = "task-1352900" [ 997.729606] env[62914]: _type = "Task" [ 997.729606] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.740427] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352900, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.759142] env[62914]: DEBUG nova.network.neutron [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Updating instance_info_cache with network_info: [{"id": "9c9a03a9-3a96-4060-97c8-3fdf53f21417", "address": "fa:16:3e:e1:35:0b", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c9a03a9-3a", "ovs_interfaceid": "9c9a03a9-3a96-4060-97c8-3fdf53f21417", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.911071] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352899, 'name': ReconfigVM_Task, 'duration_secs': 0.261556} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.911866] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4/054f37ad-4161-4c70-aa1b-b2e69cd1e9f4.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.912506] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36a97f81-8693-47a0-a59f-8722e8488bbf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.920100] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 997.920100] env[62914]: value = "task-1352901" [ 997.920100] env[62914]: _type = "Task" [ 997.920100] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.933094] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352901, 'name': Rename_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.943198] env[62914]: DEBUG oslo_vmware.api [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Task: {'id': task-1352883, 'name': ReconfigVM_Task, 'duration_secs': 6.822792} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.943456] env[62914]: DEBUG oslo_concurrency.lockutils [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] Releasing lock "1e527707-758c-4006-af83-9c739b9645ed" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.943665] env[62914]: DEBUG nova.virt.vmwareapi.vmops [req-eb703e88-e30b-464e-8d44-72e71410c4ec req-2429ec00-0c78-454a-8d71-656ebb76d085 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Reconfigured VM to detach interface {{(pid=62914) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 997.944298] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "1e527707-758c-4006-af83-9c739b9645ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 7.028s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.944541] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "1e527707-758c-4006-af83-9c739b9645ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.944849] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "1e527707-758c-4006-af83-9c739b9645ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.944919] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "1e527707-758c-4006-af83-9c739b9645ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.947159] env[62914]: INFO nova.compute.manager [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Terminating instance [ 997.949101] env[62914]: DEBUG nova.compute.manager [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 997.949305] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 997.950373] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7979cde9-d2b2-4355-bbcf-9d64dd188cd3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.956952] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 997.957199] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d5b8171-651f-4931-a137-05c370ac24d6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.964031] env[62914]: DEBUG oslo_vmware.api [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 997.964031] env[62914]: value = "task-1352902" [ 997.964031] env[62914]: _type = "Task" [ 997.964031] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.971414] env[62914]: DEBUG oslo_vmware.api [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352902, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.239302] env[62914]: DEBUG oslo_vmware.api [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352900, 'name': PowerOnVM_Task, 'duration_secs': 0.50544} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.239563] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 998.239768] env[62914]: INFO nova.compute.manager [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Took 8.33 seconds to spawn the instance on the hypervisor. [ 998.239948] env[62914]: DEBUG nova.compute.manager [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.240716] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c40854a0-21bb-45f3-b0ed-35e18a337be6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.261815] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "refresh_cache-961280ed-0160-4c4e-bca4-c5f2091e9bb0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.262108] env[62914]: DEBUG nova.compute.manager [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Instance network_info: |[{"id": "9c9a03a9-3a96-4060-97c8-3fdf53f21417", "address": "fa:16:3e:e1:35:0b", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c9a03a9-3a", "ovs_interfaceid": "9c9a03a9-3a96-4060-97c8-3fdf53f21417", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 998.262507] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:35:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '32faf59b-014c-4f1f-8331-40df95bf741f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9c9a03a9-3a96-4060-97c8-3fdf53f21417', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 998.276858] env[62914]: DEBUG oslo.service.loopingcall [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 998.276858] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 998.276858] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e21306e-2083-42ed-8bd7-98c89234c6c5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.297545] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 998.297545] env[62914]: value = "task-1352903" [ 998.297545] env[62914]: _type = "Task" [ 998.297545] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.305524] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352903, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.400481] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "2575bc0d-cb47-4dab-a219-f84254bda47d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.400893] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "2575bc0d-cb47-4dab-a219-f84254bda47d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.401293] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "2575bc0d-cb47-4dab-a219-f84254bda47d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.401426] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "2575bc0d-cb47-4dab-a219-f84254bda47d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.401608] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "2575bc0d-cb47-4dab-a219-f84254bda47d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.404167] env[62914]: INFO nova.compute.manager [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Terminating instance [ 998.405983] env[62914]: DEBUG nova.compute.manager [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 998.406255] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 998.406990] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b9f8fb-5e24-49ae-9512-a1ce8a3e8698 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.414939] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 998.415198] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-36aad357-6673-4483-bd01-7a3f7374dbde {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.421343] env[62914]: DEBUG oslo_vmware.api [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 998.421343] env[62914]: value = "task-1352904" [ 998.421343] env[62914]: _type = "Task" [ 998.421343] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.434402] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352901, 'name': Rename_Task, 'duration_secs': 0.168681} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.437551] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 998.439499] env[62914]: DEBUG oslo_vmware.api [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352904, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.439499] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-28d79a25-561f-4499-a099-410353ae536d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.443489] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 998.443489] env[62914]: value = "task-1352905" [ 998.443489] env[62914]: _type = "Task" [ 998.443489] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.451095] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352905, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.472697] env[62914]: DEBUG oslo_vmware.api [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352902, 'name': PowerOffVM_Task, 'duration_secs': 0.188798} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.472901] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 998.474729] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 998.474729] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0559c671-815a-4139-9135-fe9dffe8da82 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.514280] env[62914]: DEBUG nova.compute.manager [req-88dee552-adeb-4906-a452-19128ce6e56a req-8b6cf876-7ff1-4673-963d-ebeb018adbfa service nova] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Received event network-changed-9c9a03a9-3a96-4060-97c8-3fdf53f21417 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 998.514570] env[62914]: DEBUG nova.compute.manager [req-88dee552-adeb-4906-a452-19128ce6e56a req-8b6cf876-7ff1-4673-963d-ebeb018adbfa service nova] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Refreshing instance network info cache due to event network-changed-9c9a03a9-3a96-4060-97c8-3fdf53f21417. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 998.514968] env[62914]: DEBUG oslo_concurrency.lockutils [req-88dee552-adeb-4906-a452-19128ce6e56a req-8b6cf876-7ff1-4673-963d-ebeb018adbfa service nova] Acquiring lock "refresh_cache-961280ed-0160-4c4e-bca4-c5f2091e9bb0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.514968] env[62914]: DEBUG oslo_concurrency.lockutils [req-88dee552-adeb-4906-a452-19128ce6e56a req-8b6cf876-7ff1-4673-963d-ebeb018adbfa service nova] Acquired lock "refresh_cache-961280ed-0160-4c4e-bca4-c5f2091e9bb0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.515128] env[62914]: DEBUG nova.network.neutron [req-88dee552-adeb-4906-a452-19128ce6e56a req-8b6cf876-7ff1-4673-963d-ebeb018adbfa service nova] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Refreshing network info cache for port 9c9a03a9-3a96-4060-97c8-3fdf53f21417 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 998.538291] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 998.538712] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 998.539044] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Deleting the datastore file [datastore1] 1e527707-758c-4006-af83-9c739b9645ed {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.539435] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04e6f262-ae70-4f59-9f8a-a5e781ab56e6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.548049] env[62914]: DEBUG oslo_vmware.api [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 998.548049] env[62914]: value = "task-1352907" [ 998.548049] env[62914]: _type = "Task" [ 998.548049] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.558106] env[62914]: DEBUG oslo_vmware.api [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352907, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.762747] env[62914]: INFO nova.compute.manager [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Took 14.56 seconds to build instance. [ 998.806935] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352903, 'name': CreateVM_Task, 'duration_secs': 0.400796} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.807128] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 998.807805] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.807978] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.808327] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 998.808576] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04bff2f6-4d65-443b-9c1e-24faa89c4fbd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.815357] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 998.815357] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5255d0d8-4dc5-354f-364e-89c2bbf8af04" [ 998.815357] env[62914]: _type = "Task" [ 998.815357] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.822844] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5255d0d8-4dc5-354f-364e-89c2bbf8af04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.933980] env[62914]: DEBUG oslo_vmware.api [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352904, 'name': PowerOffVM_Task, 'duration_secs': 0.168646} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.934291] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 998.934475] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 998.934737] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2cfb2aef-5c15-4476-8ea1-9aa3cd2c2380 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.953753] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352905, 'name': PowerOnVM_Task} progress is 92%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.043471] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 999.043692] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 999.044036] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Deleting the datastore file [datastore1] 2575bc0d-cb47-4dab-a219-f84254bda47d {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 999.044383] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-919c7006-5437-48a6-979c-c6c73048d76e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.057361] env[62914]: DEBUG oslo_vmware.api [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for the task: (returnval){ [ 999.057361] env[62914]: value = "task-1352909" [ 999.057361] env[62914]: _type = "Task" [ 999.057361] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.068424] env[62914]: DEBUG oslo_vmware.api [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352907, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19437} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.070044] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.070305] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 999.070540] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 999.070734] env[62914]: INFO nova.compute.manager [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Took 1.12 seconds to destroy the instance on the hypervisor. [ 999.070981] env[62914]: DEBUG oslo.service.loopingcall [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.074695] env[62914]: DEBUG nova.compute.manager [-] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 999.074794] env[62914]: DEBUG nova.network.neutron [-] [instance: 1e527707-758c-4006-af83-9c739b9645ed] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 999.076535] env[62914]: DEBUG oslo_vmware.api [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352909, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.088062] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed63ac5f-b041-4c9e-9832-2afae5aef297 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.097170] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-49d6eeec-6818-4a20-97e5-9e7c25d9e212 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Suspending the VM {{(pid=62914) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 999.097428] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-c7f131ea-09f4-46f7-a411-59a5c5fe98e2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.103689] env[62914]: DEBUG oslo_vmware.api [None req-49d6eeec-6818-4a20-97e5-9e7c25d9e212 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 999.103689] env[62914]: value = "task-1352910" [ 999.103689] env[62914]: _type = "Task" [ 999.103689] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.111400] env[62914]: DEBUG oslo_vmware.api [None req-49d6eeec-6818-4a20-97e5-9e7c25d9e212 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352910, 'name': SuspendVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.268000] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b746a86e-b9d0-4880-9e45-c14fac0ebc01 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "84a7822d-d46d-4147-ad4c-b9402aa072f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.070s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.297095] env[62914]: DEBUG nova.network.neutron [req-88dee552-adeb-4906-a452-19128ce6e56a req-8b6cf876-7ff1-4673-963d-ebeb018adbfa service nova] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Updated VIF entry in instance network info cache for port 9c9a03a9-3a96-4060-97c8-3fdf53f21417. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 999.297717] env[62914]: DEBUG nova.network.neutron [req-88dee552-adeb-4906-a452-19128ce6e56a req-8b6cf876-7ff1-4673-963d-ebeb018adbfa service nova] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Updating instance_info_cache with network_info: [{"id": "9c9a03a9-3a96-4060-97c8-3fdf53f21417", "address": "fa:16:3e:e1:35:0b", "network": {"id": "c68c9480-c6a1-472a-9235-be88ac2c7ba7", "bridge": "br-int", "label": "tempest-ServersTestJSON-1895713608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9278818cebd842669b802a4e43e91774", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c9a03a9-3a", "ovs_interfaceid": "9c9a03a9-3a96-4060-97c8-3fdf53f21417", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.327202] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5255d0d8-4dc5-354f-364e-89c2bbf8af04, 'name': SearchDatastore_Task, 'duration_secs': 0.015925} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.327643] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.327978] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 999.328336] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.328541] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.328827] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 999.329276] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4620899a-f55b-4b91-9197-e6e4e920d7fe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.344708] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 999.345131] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 999.346320] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf7c7e91-0472-456a-9b48-94cfeae80896 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.352591] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 999.352591] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5279f574-a4cc-a93e-7030-ee30c0442a2d" [ 999.352591] env[62914]: _type = "Task" [ 999.352591] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.361094] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5279f574-a4cc-a93e-7030-ee30c0442a2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.454567] env[62914]: DEBUG oslo_vmware.api [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352905, 'name': PowerOnVM_Task, 'duration_secs': 0.552928} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.454926] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 999.455171] env[62914]: INFO nova.compute.manager [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Took 7.29 seconds to spawn the instance on the hypervisor. [ 999.455366] env[62914]: DEBUG nova.compute.manager [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 999.456326] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b582226a-d464-42ce-a3d7-e21d5ef70783 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.567190] env[62914]: DEBUG oslo_vmware.api [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Task: {'id': task-1352909, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.422684} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.567465] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.567659] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 999.567875] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 999.568082] env[62914]: INFO nova.compute.manager [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 999.568352] env[62914]: DEBUG oslo.service.loopingcall [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.569303] env[62914]: DEBUG nova.compute.manager [-] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 999.569303] env[62914]: DEBUG nova.network.neutron [-] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 999.613644] env[62914]: DEBUG oslo_vmware.api [None req-49d6eeec-6818-4a20-97e5-9e7c25d9e212 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352910, 'name': SuspendVM_Task} progress is 58%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.801632] env[62914]: DEBUG oslo_concurrency.lockutils [req-88dee552-adeb-4906-a452-19128ce6e56a req-8b6cf876-7ff1-4673-963d-ebeb018adbfa service nova] Releasing lock "refresh_cache-961280ed-0160-4c4e-bca4-c5f2091e9bb0" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.827137] env[62914]: DEBUG nova.compute.manager [req-6b37f558-8440-4ac8-9ad2-c81dfd5ca4b1 req-f2d05051-2f78-412f-b030-6a4d8ed6c1a1 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Received event network-vif-deleted-615c8b1c-438d-43fb-999b-89172ade3f61 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 999.827664] env[62914]: INFO nova.compute.manager [req-6b37f558-8440-4ac8-9ad2-c81dfd5ca4b1 req-f2d05051-2f78-412f-b030-6a4d8ed6c1a1 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Neutron deleted interface 615c8b1c-438d-43fb-999b-89172ade3f61; detaching it from the instance and deleting it from the info cache [ 999.827664] env[62914]: DEBUG nova.network.neutron [req-6b37f558-8440-4ac8-9ad2-c81dfd5ca4b1 req-f2d05051-2f78-412f-b030-6a4d8ed6c1a1 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.862556] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5279f574-a4cc-a93e-7030-ee30c0442a2d, 'name': SearchDatastore_Task, 'duration_secs': 0.058334} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.863452] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-424c865a-913f-4630-8356-2533dce90a7c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.869387] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 999.869387] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52323e13-24a7-ca53-8980-f66ed4ec258f" [ 999.869387] env[62914]: _type = "Task" [ 999.869387] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.878195] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52323e13-24a7-ca53-8980-f66ed4ec258f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.975797] env[62914]: INFO nova.compute.manager [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Took 14.47 seconds to build instance. [ 1000.114447] env[62914]: DEBUG oslo_vmware.api [None req-49d6eeec-6818-4a20-97e5-9e7c25d9e212 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352910, 'name': SuspendVM_Task, 'duration_secs': 0.879977} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.115273] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-49d6eeec-6818-4a20-97e5-9e7c25d9e212 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Suspended the VM {{(pid=62914) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1000.115273] env[62914]: DEBUG nova.compute.manager [None req-49d6eeec-6818-4a20-97e5-9e7c25d9e212 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1000.115956] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f74b42-3aa4-4936-93c7-4f5a243f746c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.274269] env[62914]: DEBUG nova.network.neutron [-] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.334018] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4c5915eb-17ab-474b-ad83-66926645b8cd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.343946] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9543b4ce-8698-4c2a-900d-b39d02031f52 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.373592] env[62914]: DEBUG nova.compute.manager [req-6b37f558-8440-4ac8-9ad2-c81dfd5ca4b1 req-f2d05051-2f78-412f-b030-6a4d8ed6c1a1 service nova] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Detach interface failed, port_id=615c8b1c-438d-43fb-999b-89172ade3f61, reason: Instance 1e527707-758c-4006-af83-9c739b9645ed could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1000.382512] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52323e13-24a7-ca53-8980-f66ed4ec258f, 'name': SearchDatastore_Task, 'duration_secs': 0.010871} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.382782] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.383083] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 961280ed-0160-4c4e-bca4-c5f2091e9bb0/961280ed-0160-4c4e-bca4-c5f2091e9bb0.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1000.383346] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-83f97cb7-4f24-4339-831c-faa12dec86d9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.390656] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 1000.390656] env[62914]: value = "task-1352911" [ 1000.390656] env[62914]: _type = "Task" [ 1000.390656] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.398200] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352911, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.479626] env[62914]: DEBUG oslo_concurrency.lockutils [None req-baa49ba8-dc14-4504-bb40-fa48f4718402 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "054f37ad-4161-4c70-aa1b-b2e69cd1e9f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.979s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.549471] env[62914]: DEBUG nova.compute.manager [req-8f497b4a-052c-4de8-a539-1f62a9a83e95 req-97000a2d-b104-4119-97d4-be92fb8c3d7e service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Received event network-vif-deleted-d41f0fe0-d70c-4f8c-a306-e14dc0107565 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.549590] env[62914]: INFO nova.compute.manager [req-8f497b4a-052c-4de8-a539-1f62a9a83e95 req-97000a2d-b104-4119-97d4-be92fb8c3d7e service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Neutron deleted interface d41f0fe0-d70c-4f8c-a306-e14dc0107565; detaching it from the instance and deleting it from the info cache [ 1000.549725] env[62914]: DEBUG nova.network.neutron [req-8f497b4a-052c-4de8-a539-1f62a9a83e95 req-97000a2d-b104-4119-97d4-be92fb8c3d7e service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.594110] env[62914]: DEBUG nova.network.neutron [-] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.777269] env[62914]: INFO nova.compute.manager [-] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Took 1.70 seconds to deallocate network for instance. [ 1000.901365] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352911, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.053953] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67846135-a966-457a-82b1-7d3346294a94 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.063255] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d224cd55-7456-4e58-96e8-70a01a7f82a3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.096320] env[62914]: INFO nova.compute.manager [-] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Took 1.53 seconds to deallocate network for instance. [ 1001.096742] env[62914]: DEBUG nova.compute.manager [req-8f497b4a-052c-4de8-a539-1f62a9a83e95 req-97000a2d-b104-4119-97d4-be92fb8c3d7e service nova] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Detach interface failed, port_id=d41f0fe0-d70c-4f8c-a306-e14dc0107565, reason: Instance 2575bc0d-cb47-4dab-a219-f84254bda47d could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1001.284731] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.284731] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.284840] env[62914]: DEBUG nova.objects.instance [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'resources' on Instance uuid 1e527707-758c-4006-af83-9c739b9645ed {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.401526] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352911, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.932992} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.401797] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 961280ed-0160-4c4e-bca4-c5f2091e9bb0/961280ed-0160-4c4e-bca4-c5f2091e9bb0.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1001.402079] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1001.402348] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6fe4148a-7a88-441f-9c8d-6cda2b39529a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.408443] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 1001.408443] env[62914]: value = "task-1352912" [ 1001.408443] env[62914]: _type = "Task" [ 1001.408443] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.415942] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352912, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.582220] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "84a7822d-d46d-4147-ad4c-b9402aa072f4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.582611] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "84a7822d-d46d-4147-ad4c-b9402aa072f4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.582716] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "84a7822d-d46d-4147-ad4c-b9402aa072f4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.582948] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "84a7822d-d46d-4147-ad4c-b9402aa072f4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.583167] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "84a7822d-d46d-4147-ad4c-b9402aa072f4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.585529] env[62914]: INFO nova.compute.manager [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Terminating instance [ 1001.587397] env[62914]: DEBUG nova.compute.manager [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1001.587611] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1001.588494] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04915854-dbf0-432f-af5c-ce5ea47e7c0b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.596221] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1001.596485] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8f90e5ff-a63f-43a5-9454-6e72004eebfd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.604292] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.666635] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.667071] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.667071] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleting the datastore file [datastore2] 84a7822d-d46d-4147-ad4c-b9402aa072f4 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.667353] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4fbdad80-5412-4bfd-ae70-1e1ca658c23b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.673643] env[62914]: DEBUG oslo_vmware.api [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1001.673643] env[62914]: value = "task-1352914" [ 1001.673643] env[62914]: _type = "Task" [ 1001.673643] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.682256] env[62914]: DEBUG oslo_vmware.api [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352914, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.879517] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61acf31-a491-459d-ad82-73a4c39b6d82 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.887031] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64e7a5f-2722-4d90-89fd-9a1d3d90a680 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.923392] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0befc6d3-7bc5-40f3-9f84-ecac9f5ba463 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.933393] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e33b53-c18c-4b0a-8001-eed22e72cf17 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.937139] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352912, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06572} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.937414] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1001.938497] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eabfa54e-e3d9-4540-a9a2-744d759e55f0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.948835] env[62914]: DEBUG nova.compute.provider_tree [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.969515] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 961280ed-0160-4c4e-bca4-c5f2091e9bb0/961280ed-0160-4c4e-bca4-c5f2091e9bb0.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.970576] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff8d3abb-7521-4861-ac49-37f857406429 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.989904] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 1001.989904] env[62914]: value = "task-1352915" [ 1001.989904] env[62914]: _type = "Task" [ 1001.989904] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.998394] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352915, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.153077] env[62914]: DEBUG nova.compute.manager [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.154045] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9576a50c-3d35-49a6-b238-fb3a92a8a52a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.184783] env[62914]: DEBUG oslo_vmware.api [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352914, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.285641} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.184783] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.185012] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.185012] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.185226] env[62914]: INFO nova.compute.manager [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1002.185511] env[62914]: DEBUG oslo.service.loopingcall [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.185715] env[62914]: DEBUG nova.compute.manager [-] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1002.185812] env[62914]: DEBUG nova.network.neutron [-] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1002.452815] env[62914]: DEBUG nova.scheduler.client.report [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1002.499999] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.573393] env[62914]: DEBUG nova.compute.manager [req-8cffd4c4-6bf7-4fa6-8e62-86cfb6bd820d req-da4f67db-fa9a-49ed-a214-bd387abf3795 service nova] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Received event network-vif-deleted-d04462d2-47dc-4db8-838d-0912b80a1db3 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.573598] env[62914]: INFO nova.compute.manager [req-8cffd4c4-6bf7-4fa6-8e62-86cfb6bd820d req-da4f67db-fa9a-49ed-a214-bd387abf3795 service nova] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Neutron deleted interface d04462d2-47dc-4db8-838d-0912b80a1db3; detaching it from the instance and deleting it from the info cache [ 1002.573773] env[62914]: DEBUG nova.network.neutron [req-8cffd4c4-6bf7-4fa6-8e62-86cfb6bd820d req-da4f67db-fa9a-49ed-a214-bd387abf3795 service nova] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.667959] env[62914]: INFO nova.compute.manager [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] instance snapshotting [ 1002.670936] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf24132-00cf-49fc-b5f6-4dbe30a19c04 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.689463] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb93d51-a973-4f5b-9d9d-9b23d54fa73e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.961519] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.677s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.964676] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.360s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.965200] env[62914]: DEBUG nova.objects.instance [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lazy-loading 'resources' on Instance uuid 2575bc0d-cb47-4dab-a219-f84254bda47d {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.984296] env[62914]: INFO nova.scheduler.client.report [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Deleted allocations for instance 1e527707-758c-4006-af83-9c739b9645ed [ 1002.989601] env[62914]: DEBUG nova.network.neutron [-] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.999582] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352915, 'name': ReconfigVM_Task, 'duration_secs': 0.931033} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.000486] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 961280ed-0160-4c4e-bca4-c5f2091e9bb0/961280ed-0160-4c4e-bca4-c5f2091e9bb0.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.001202] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ba389f2-8c37-43ba-81ab-e3e6e60d2223 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.008657] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 1003.008657] env[62914]: value = "task-1352916" [ 1003.008657] env[62914]: _type = "Task" [ 1003.008657] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.016921] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352916, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.078294] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21fa8e55-f972-4ca3-a889-6a5728089f69 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.087952] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0750de44-a9d6-49e4-9389-191a64c445fa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.113244] env[62914]: DEBUG nova.compute.manager [req-8cffd4c4-6bf7-4fa6-8e62-86cfb6bd820d req-da4f67db-fa9a-49ed-a214-bd387abf3795 service nova] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Detach interface failed, port_id=d04462d2-47dc-4db8-838d-0912b80a1db3, reason: Instance 84a7822d-d46d-4147-ad4c-b9402aa072f4 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1003.199856] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Creating Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1003.200370] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-eddb5ebb-939e-47cd-922f-f71ffc03e87a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.207169] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 1003.207169] env[62914]: value = "task-1352917" [ 1003.207169] env[62914]: _type = "Task" [ 1003.207169] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.215630] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352917, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.494286] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb6e453b-f400-4b5d-8313-e203f22a4722 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "1e527707-758c-4006-af83-9c739b9645ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.550s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.495317] env[62914]: INFO nova.compute.manager [-] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Took 1.31 seconds to deallocate network for instance. [ 1003.524323] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352916, 'name': Rename_Task, 'duration_secs': 0.222288} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.524666] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1003.524936] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b203f7f-23da-4e19-8aba-3a5e73f39d90 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.536078] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 1003.536078] env[62914]: value = "task-1352918" [ 1003.536078] env[62914]: _type = "Task" [ 1003.536078] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.551748] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352918, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.569568] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16797bbe-b41d-4938-8b1f-fda846d76fb2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.578030] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8828d5-e6ad-4c01-9f6b-e3e37ba6159d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.613571] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb06d3a-a27a-4109-b330-e55c76b3813e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.622093] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c910e58-1ba2-4c7e-858d-aa3ed77dc123 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.638158] env[62914]: DEBUG nova.compute.provider_tree [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1003.719488] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352917, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.727325] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.727704] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.003359] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.045313] env[62914]: DEBUG oslo_vmware.api [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352918, 'name': PowerOnVM_Task, 'duration_secs': 0.490861} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.045603] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1004.045814] env[62914]: INFO nova.compute.manager [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Took 7.98 seconds to spawn the instance on the hypervisor. [ 1004.045998] env[62914]: DEBUG nova.compute.manager [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1004.046784] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943eb2e5-d167-4f68-9e6b-37248d9dc37d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.144790] env[62914]: DEBUG nova.scheduler.client.report [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1004.218474] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352917, 'name': CreateSnapshot_Task, 'duration_secs': 0.577047} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.218761] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Created Snapshot of the VM instance {{(pid=62914) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1004.219505] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662fd3ed-d2f4-4b51-99eb-24f7f59130da {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.230656] env[62914]: DEBUG nova.compute.utils [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1004.566918] env[62914]: INFO nova.compute.manager [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Took 16.68 seconds to build instance. [ 1004.648558] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.684s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.650983] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.648s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.651298] env[62914]: DEBUG nova.objects.instance [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lazy-loading 'resources' on Instance uuid 84a7822d-d46d-4147-ad4c-b9402aa072f4 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1004.668129] env[62914]: INFO nova.scheduler.client.report [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Deleted allocations for instance 2575bc0d-cb47-4dab-a219-f84254bda47d [ 1004.738911] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Creating linked-clone VM from snapshot {{(pid=62914) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1004.739559] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.740060] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-41cf5cc6-c7f9-4fdf-ba06-44da231e6d7b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.749816] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 1004.749816] env[62914]: value = "task-1352919" [ 1004.749816] env[62914]: _type = "Task" [ 1004.749816] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.757762] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352919, 'name': CloneVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.064158] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.064449] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.068598] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7cb82d94-6362-4ddb-a7d5-229392c89551 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.198s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.177207] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5ace04b6-58b4-47fa-a7c7-c0f795123871 tempest-AttachVolumeNegativeTest-916243774 tempest-AttachVolumeNegativeTest-916243774-project-member] Lock "2575bc0d-cb47-4dab-a219-f84254bda47d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.776s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.235365] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffe6641d-bc4e-4750-943f-cb004d978659 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.244589] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1724a94c-d2f9-4283-aa31-4c537709e1ae {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.278498] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70250ab2-35b6-404a-a5f7-f600f8dd9063 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.284359] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352919, 'name': CloneVM_Task} progress is 94%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.289145] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ffab9fc-95af-45f4-b078-4ac0ba7b8a0a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.302800] env[62914]: DEBUG nova.compute.provider_tree [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.536290] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e93b43b4-4a59-4450-a4b2-dfc3edd0ead3 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.536497] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e93b43b4-4a59-4450-a4b2-dfc3edd0ead3 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.536681] env[62914]: DEBUG nova.compute.manager [None req-e93b43b4-4a59-4450-a4b2-dfc3edd0ead3 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1005.537589] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9c79b4-dbc8-43f1-9f69-1f91029b0337 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.544975] env[62914]: DEBUG nova.compute.manager [None req-e93b43b4-4a59-4450-a4b2-dfc3edd0ead3 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62914) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1005.545578] env[62914]: DEBUG nova.objects.instance [None req-e93b43b4-4a59-4450-a4b2-dfc3edd0ead3 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lazy-loading 'flavor' on Instance uuid 961280ed-0160-4c4e-bca4-c5f2091e9bb0 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.567080] env[62914]: DEBUG nova.compute.manager [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1005.760315] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352919, 'name': CloneVM_Task} progress is 95%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.808681] env[62914]: DEBUG nova.scheduler.client.report [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1005.812322] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.812559] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.812783] env[62914]: INFO nova.compute.manager [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Attaching volume 96b3ee28-51d2-440e-a63d-387302c4807d to /dev/sdb [ 1005.844546] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de24091d-088b-4f77-8050-385deccb5196 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.853538] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4aa0cd-ba22-4f36-b740-4eb507a8c780 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.866834] env[62914]: DEBUG nova.virt.block_device [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Updating existing volume attachment record: 6bafe5f1-a9e4-4ca2-8f9a-f6806c01704c {{(pid=62914) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1006.053842] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-e93b43b4-4a59-4450-a4b2-dfc3edd0ead3 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1006.053842] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10e09527-68f3-4ec7-a0a7-4edfc62c3436 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.058935] env[62914]: DEBUG oslo_vmware.api [None req-e93b43b4-4a59-4450-a4b2-dfc3edd0ead3 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 1006.058935] env[62914]: value = "task-1352922" [ 1006.058935] env[62914]: _type = "Task" [ 1006.058935] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.074150] env[62914]: DEBUG oslo_vmware.api [None req-e93b43b4-4a59-4450-a4b2-dfc3edd0ead3 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352922, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.097679] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.260578] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352919, 'name': CloneVM_Task, 'duration_secs': 1.226983} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.260879] env[62914]: INFO nova.virt.vmwareapi.vmops [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Created linked-clone VM from snapshot [ 1006.261662] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512e1fd1-7260-45d1-9250-5b69b2d95fcb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.268693] env[62914]: DEBUG nova.virt.vmwareapi.images [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Uploading image abcb1063-0eb8-42f6-8e45-5217f0daf6b0 {{(pid=62914) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1006.281935] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Destroying the VM {{(pid=62914) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1006.282224] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-19904133-0ec9-43e1-812a-d5cfa536e296 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.288314] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 1006.288314] env[62914]: value = "task-1352925" [ 1006.288314] env[62914]: _type = "Task" [ 1006.288314] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.297026] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352925, 'name': Destroy_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.315991] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.665s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.318467] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.221s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.320017] env[62914]: INFO nova.compute.claims [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1006.335101] env[62914]: INFO nova.scheduler.client.report [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleted allocations for instance 84a7822d-d46d-4147-ad4c-b9402aa072f4 [ 1006.569138] env[62914]: DEBUG oslo_vmware.api [None req-e93b43b4-4a59-4450-a4b2-dfc3edd0ead3 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352922, 'name': PowerOffVM_Task, 'duration_secs': 0.183922} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.569370] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-e93b43b4-4a59-4450-a4b2-dfc3edd0ead3 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1006.569549] env[62914]: DEBUG nova.compute.manager [None req-e93b43b4-4a59-4450-a4b2-dfc3edd0ead3 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.570361] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7910ad6-eee1-4c48-9ce4-4fbcad07f53b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.797974] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352925, 'name': Destroy_Task, 'duration_secs': 0.408566} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.798982] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Destroyed the VM [ 1006.799104] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Deleting Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1006.800040] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-89ee68a4-b892-49dd-9415-ca32f16b6c3d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.805487] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 1006.805487] env[62914]: value = "task-1352926" [ 1006.805487] env[62914]: _type = "Task" [ 1006.805487] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.812926] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352926, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.845499] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6e6c0b1f-257c-4710-b167-8b731b23b133 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "84a7822d-d46d-4147-ad4c-b9402aa072f4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.263s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.082054] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e93b43b4-4a59-4450-a4b2-dfc3edd0ead3 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.545s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.315979] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352926, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.396762] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53cb22aa-5aa0-4bcb-8608-5f61bd306e6c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.406406] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18d5669c-749e-438b-8cff-77374f98d41d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.436813] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-065c7702-0a02-430f-8d25-6ecd403866d9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.444031] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f2e54b-75eb-47e6-a85d-f757602ff959 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.457330] env[62914]: DEBUG nova.compute.provider_tree [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.616562] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "824f38d6-26c0-4922-ad3d-111b21bf08ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.616808] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "824f38d6-26c0-4922-ad3d-111b21bf08ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.815614] env[62914]: DEBUG oslo_vmware.api [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352926, 'name': RemoveSnapshot_Task, 'duration_secs': 0.719504} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.815899] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Deleted Snapshot of the VM instance {{(pid=62914) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1007.960951] env[62914]: DEBUG nova.scheduler.client.report [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.118782] env[62914]: DEBUG nova.compute.manager [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1008.321955] env[62914]: WARNING nova.compute.manager [None req-d84c18ef-7b9d-48f9-877d-c22b2e2b8454 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Image not found during snapshot: nova.exception.ImageNotFound: Image abcb1063-0eb8-42f6-8e45-5217f0daf6b0 could not be found. [ 1008.363069] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.363319] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.363614] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.363984] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.364346] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.366365] env[62914]: INFO nova.compute.manager [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Terminating instance [ 1008.368074] env[62914]: DEBUG nova.compute.manager [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1008.368328] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1008.369176] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa6d9af-968b-432f-9f4a-08cc1a253140 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.376778] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1008.377018] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-54607fa5-047f-4708-b5f8-c3d211478ca1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.466076] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.147s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.466829] env[62914]: DEBUG nova.compute.manager [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1008.642449] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.642717] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.644287] env[62914]: INFO nova.compute.claims [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1008.840134] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "054f37ad-4161-4c70-aa1b-b2e69cd1e9f4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.840422] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "054f37ad-4161-4c70-aa1b-b2e69cd1e9f4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.840638] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "054f37ad-4161-4c70-aa1b-b2e69cd1e9f4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.840826] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "054f37ad-4161-4c70-aa1b-b2e69cd1e9f4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.841011] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "054f37ad-4161-4c70-aa1b-b2e69cd1e9f4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.843225] env[62914]: INFO nova.compute.manager [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Terminating instance [ 1008.845129] env[62914]: DEBUG nova.compute.manager [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1008.845338] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1008.846171] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51758ea-2d17-40fa-8c72-ffa202e18d43 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.854104] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1008.854350] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d831b5b-a6c8-4e73-b1b6-fbcb3f44d96d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.860436] env[62914]: DEBUG oslo_vmware.api [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 1008.860436] env[62914]: value = "task-1352929" [ 1008.860436] env[62914]: _type = "Task" [ 1008.860436] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.867735] env[62914]: DEBUG oslo_vmware.api [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352929, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.938697] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1008.938920] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1008.939125] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleting the datastore file [datastore2] 961280ed-0160-4c4e-bca4-c5f2091e9bb0 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1008.939401] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8640d9cd-5e71-4521-bac9-3fa465b6a542 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.946239] env[62914]: DEBUG oslo_vmware.api [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for the task: (returnval){ [ 1008.946239] env[62914]: value = "task-1352930" [ 1008.946239] env[62914]: _type = "Task" [ 1008.946239] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.957943] env[62914]: DEBUG oslo_vmware.api [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352930, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.973175] env[62914]: DEBUG nova.compute.utils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1008.975515] env[62914]: DEBUG nova.compute.manager [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1008.975515] env[62914]: DEBUG nova.network.neutron [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1009.028510] env[62914]: DEBUG nova.policy [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80089586cfda4cd2bdfc24d9f556239c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09d51090ac5d41fc966c7e132f288824', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 1009.298688] env[62914]: DEBUG nova.network.neutron [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Successfully created port: c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1009.371399] env[62914]: DEBUG oslo_vmware.api [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352929, 'name': PowerOffVM_Task, 'duration_secs': 0.222223} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.371762] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1009.372012] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1009.372335] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-469242c3-e004-4bc4-9632-4ebff6a3db31 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.443491] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1009.443716] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1009.443891] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleting the datastore file [datastore2] 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1009.444199] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-08d3c02e-0a9a-43d9-b165-9cba1c75ced4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.457167] env[62914]: DEBUG oslo_vmware.api [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for the task: (returnval){ [ 1009.457167] env[62914]: value = "task-1352933" [ 1009.457167] env[62914]: _type = "Task" [ 1009.457167] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.460152] env[62914]: DEBUG oslo_vmware.api [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Task: {'id': task-1352930, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144132} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.463493] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.463776] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1009.463996] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1009.464228] env[62914]: INFO nova.compute.manager [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1009.464499] env[62914]: DEBUG oslo.service.loopingcall [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.464732] env[62914]: DEBUG nova.compute.manager [-] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1009.464855] env[62914]: DEBUG nova.network.neutron [-] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1009.470927] env[62914]: DEBUG oslo_vmware.api [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352933, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.479535] env[62914]: DEBUG nova.compute.manager [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1009.728339] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f992c3-81ed-4462-a189-57370bbc5955 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.735760] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b1e3df-724a-4608-a3ac-78190b3ca2a4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.767146] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e048e53-5416-41a5-9269-8d7cb55e7141 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.771755] env[62914]: DEBUG nova.compute.manager [req-96f9e00f-5b91-4b7f-9017-a48de6b65978 req-78d355f6-045b-453a-b66d-589da0a624f6 service nova] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Received event network-vif-deleted-9c9a03a9-3a96-4060-97c8-3fdf53f21417 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.771942] env[62914]: INFO nova.compute.manager [req-96f9e00f-5b91-4b7f-9017-a48de6b65978 req-78d355f6-045b-453a-b66d-589da0a624f6 service nova] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Neutron deleted interface 9c9a03a9-3a96-4060-97c8-3fdf53f21417; detaching it from the instance and deleting it from the info cache [ 1009.772130] env[62914]: DEBUG nova.network.neutron [req-96f9e00f-5b91-4b7f-9017-a48de6b65978 req-78d355f6-045b-453a-b66d-589da0a624f6 service nova] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.778111] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e9737d-f2c5-41b7-8003-b91c4e14ab64 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.791618] env[62914]: DEBUG nova.compute.provider_tree [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.971045] env[62914]: DEBUG oslo_vmware.api [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Task: {'id': task-1352933, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156978} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.971394] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.971447] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1009.971612] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1009.971791] env[62914]: INFO nova.compute.manager [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1009.972040] env[62914]: DEBUG oslo.service.loopingcall [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.972241] env[62914]: DEBUG nova.compute.manager [-] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1009.972331] env[62914]: DEBUG nova.network.neutron [-] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1010.266471] env[62914]: DEBUG nova.network.neutron [-] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.274843] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d2025501-8b35-42f7-a4f0-6ffe66879dd6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.285292] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef09864e-1ee3-454f-a4ad-afb20b806b9e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.296741] env[62914]: DEBUG nova.scheduler.client.report [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.313367] env[62914]: DEBUG nova.compute.manager [req-96f9e00f-5b91-4b7f-9017-a48de6b65978 req-78d355f6-045b-453a-b66d-589da0a624f6 service nova] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Detach interface failed, port_id=9c9a03a9-3a96-4060-97c8-3fdf53f21417, reason: Instance 961280ed-0160-4c4e-bca4-c5f2091e9bb0 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1010.414809] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Volume attach. Driver type: vmdk {{(pid=62914) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1010.415174] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288265', 'volume_id': '96b3ee28-51d2-440e-a63d-387302c4807d', 'name': 'volume-96b3ee28-51d2-440e-a63d-387302c4807d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '062cd4d7-3cf1-4f31-9631-e9d1e194ad53', 'attached_at': '', 'detached_at': '', 'volume_id': '96b3ee28-51d2-440e-a63d-387302c4807d', 'serial': '96b3ee28-51d2-440e-a63d-387302c4807d'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1010.416150] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8784d4-b49d-4469-ba14-4737679758e7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.432083] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ce1f18-284e-4d71-b287-33c28a943574 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.456258] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] volume-96b3ee28-51d2-440e-a63d-387302c4807d/volume-96b3ee28-51d2-440e-a63d-387302c4807d.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1010.456545] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8342be0b-dab6-4537-a927-12769b684d85 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.476420] env[62914]: DEBUG oslo_vmware.api [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1010.476420] env[62914]: value = "task-1352934" [ 1010.476420] env[62914]: _type = "Task" [ 1010.476420] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.486635] env[62914]: DEBUG oslo_vmware.api [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352934, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.488894] env[62914]: DEBUG nova.compute.manager [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1010.517263] env[62914]: DEBUG nova.virt.hardware [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1010.517610] env[62914]: DEBUG nova.virt.hardware [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1010.517859] env[62914]: DEBUG nova.virt.hardware [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1010.518159] env[62914]: DEBUG nova.virt.hardware [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1010.518429] env[62914]: DEBUG nova.virt.hardware [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1010.518679] env[62914]: DEBUG nova.virt.hardware [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1010.519039] env[62914]: DEBUG nova.virt.hardware [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1010.519322] env[62914]: DEBUG nova.virt.hardware [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1010.519603] env[62914]: DEBUG nova.virt.hardware [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1010.519881] env[62914]: DEBUG nova.virt.hardware [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1010.520575] env[62914]: DEBUG nova.virt.hardware [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1010.522214] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8634f1d1-c297-421c-ba45-64ef353015b7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.532207] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd3fea1-d7e2-4e28-af6b-ec660d5390a8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.245096] env[62914]: DEBUG nova.network.neutron [-] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.245477] env[62914]: INFO nova.compute.manager [-] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Took 1.78 seconds to deallocate network for instance. [ 1011.245853] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.603s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.246598] env[62914]: DEBUG nova.compute.manager [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1011.251674] env[62914]: DEBUG nova.network.neutron [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Successfully updated port: c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1011.257883] env[62914]: DEBUG nova.compute.manager [req-55ec99c9-43e4-4ea3-a09c-ba92bd6243e5 req-3fe7158e-2806-4f69-a4ea-bf44226884e2 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Received event network-vif-plugged-c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.258143] env[62914]: DEBUG oslo_concurrency.lockutils [req-55ec99c9-43e4-4ea3-a09c-ba92bd6243e5 req-3fe7158e-2806-4f69-a4ea-bf44226884e2 service nova] Acquiring lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.261927] env[62914]: DEBUG oslo_concurrency.lockutils [req-55ec99c9-43e4-4ea3-a09c-ba92bd6243e5 req-3fe7158e-2806-4f69-a4ea-bf44226884e2 service nova] Lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.261927] env[62914]: DEBUG oslo_concurrency.lockutils [req-55ec99c9-43e4-4ea3-a09c-ba92bd6243e5 req-3fe7158e-2806-4f69-a4ea-bf44226884e2 service nova] Lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.261927] env[62914]: DEBUG nova.compute.manager [req-55ec99c9-43e4-4ea3-a09c-ba92bd6243e5 req-3fe7158e-2806-4f69-a4ea-bf44226884e2 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] No waiting events found dispatching network-vif-plugged-c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1011.261927] env[62914]: WARNING nova.compute.manager [req-55ec99c9-43e4-4ea3-a09c-ba92bd6243e5 req-3fe7158e-2806-4f69-a4ea-bf44226884e2 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Received unexpected event network-vif-plugged-c240aa60-b896-4756-9de5-3965cfc8b21a for instance with vm_state building and task_state spawning. [ 1011.269575] env[62914]: DEBUG oslo_vmware.api [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352934, 'name': ReconfigVM_Task, 'duration_secs': 0.378958} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.269575] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Reconfigured VM instance instance-0000005d to attach disk [datastore1] volume-96b3ee28-51d2-440e-a63d-387302c4807d/volume-96b3ee28-51d2-440e-a63d-387302c4807d.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1011.274784] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28423887-2d2f-452d-8df4-97f6fb1641df {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.290824] env[62914]: DEBUG oslo_vmware.api [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1011.290824] env[62914]: value = "task-1352935" [ 1011.290824] env[62914]: _type = "Task" [ 1011.290824] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.300729] env[62914]: DEBUG oslo_vmware.api [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352935, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.752211] env[62914]: INFO nova.compute.manager [-] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Took 1.78 seconds to deallocate network for instance. [ 1011.754422] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.754558] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.754699] env[62914]: DEBUG nova.network.neutron [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1011.758972] env[62914]: DEBUG nova.compute.utils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1011.761552] env[62914]: DEBUG nova.compute.manager [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1011.761727] env[62914]: DEBUG nova.network.neutron [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1011.765207] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.765496] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.765711] env[62914]: DEBUG nova.objects.instance [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lazy-loading 'resources' on Instance uuid 961280ed-0160-4c4e-bca4-c5f2091e9bb0 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.797082] env[62914]: DEBUG nova.compute.manager [req-ad5c7d35-769a-41bd-919c-b71ee647c0d8 req-a94d6214-a874-4262-8ce6-7736c2721879 service nova] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Received event network-vif-deleted-fadaceea-8eb6-4ab7-9c90-8095a38ec8cf {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.803342] env[62914]: DEBUG oslo_vmware.api [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352935, 'name': ReconfigVM_Task, 'duration_secs': 0.155863} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.803664] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288265', 'volume_id': '96b3ee28-51d2-440e-a63d-387302c4807d', 'name': 'volume-96b3ee28-51d2-440e-a63d-387302c4807d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '062cd4d7-3cf1-4f31-9631-e9d1e194ad53', 'attached_at': '', 'detached_at': '', 'volume_id': '96b3ee28-51d2-440e-a63d-387302c4807d', 'serial': '96b3ee28-51d2-440e-a63d-387302c4807d'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1011.816579] env[62914]: DEBUG nova.policy [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2c21c59e39b94deea6c3701fe971198a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1dcda44f0d6a455bb17966e7c5061190', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 1012.054563] env[62914]: DEBUG nova.network.neutron [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Successfully created port: f59c63fd-416a-42f8-9c9f-1b19b5003020 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1012.284245] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.284245] env[62914]: DEBUG nova.compute.manager [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1012.286839] env[62914]: DEBUG nova.network.neutron [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1012.391292] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d5cf91-22f3-4991-b1a8-ce3dd2bca848 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.399635] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08d227a-36b1-45c6-a8aa-67413041c4f4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.434372] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d264e7e-dbdc-46ec-ab44-55995163f06c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.442080] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe7bbcb-8845-4b6a-9178-60dd49a600ac {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.455842] env[62914]: DEBUG nova.compute.provider_tree [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.705021] env[62914]: DEBUG nova.network.neutron [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updating instance_info_cache with network_info: [{"id": "c240aa60-b896-4756-9de5-3965cfc8b21a", "address": "fa:16:3e:23:5d:80", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc240aa60-b8", "ovs_interfaceid": "c240aa60-b896-4756-9de5-3965cfc8b21a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.747092] env[62914]: DEBUG nova.compute.manager [req-25b2e8a2-5cca-41ec-8952-69ea43e448ce req-d3affacc-a608-4245-aa3c-807fb25e46bd service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Received event network-changed-c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.747092] env[62914]: DEBUG nova.compute.manager [req-25b2e8a2-5cca-41ec-8952-69ea43e448ce req-d3affacc-a608-4245-aa3c-807fb25e46bd service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Refreshing instance network info cache due to event network-changed-c240aa60-b896-4756-9de5-3965cfc8b21a. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1012.747092] env[62914]: DEBUG oslo_concurrency.lockutils [req-25b2e8a2-5cca-41ec-8952-69ea43e448ce req-d3affacc-a608-4245-aa3c-807fb25e46bd service nova] Acquiring lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.873684] env[62914]: DEBUG nova.objects.instance [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lazy-loading 'flavor' on Instance uuid 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1012.959122] env[62914]: DEBUG nova.scheduler.client.report [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.166916] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a8cdf56f-2c00-40af-a806-567613f93187 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.207674] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.208047] env[62914]: DEBUG nova.compute.manager [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Instance network_info: |[{"id": "c240aa60-b896-4756-9de5-3965cfc8b21a", "address": "fa:16:3e:23:5d:80", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc240aa60-b8", "ovs_interfaceid": "c240aa60-b896-4756-9de5-3965cfc8b21a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1013.208374] env[62914]: DEBUG oslo_concurrency.lockutils [req-25b2e8a2-5cca-41ec-8952-69ea43e448ce req-d3affacc-a608-4245-aa3c-807fb25e46bd service nova] Acquired lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.208570] env[62914]: DEBUG nova.network.neutron [req-25b2e8a2-5cca-41ec-8952-69ea43e448ce req-d3affacc-a608-4245-aa3c-807fb25e46bd service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Refreshing network info cache for port c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1013.209835] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:5d:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e31264e2-3e0a-4dfb-ba1f-6389d7d47548', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c240aa60-b896-4756-9de5-3965cfc8b21a', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.217694] env[62914]: DEBUG oslo.service.loopingcall [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.220428] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1013.220893] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e29aa600-a803-406c-900d-82f8fbfe6fae {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.241677] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.241677] env[62914]: value = "task-1352936" [ 1013.241677] env[62914]: _type = "Task" [ 1013.241677] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.250924] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352936, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.277440] env[62914]: DEBUG nova.compute.manager [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1013.310203] env[62914]: DEBUG nova.virt.hardware [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.310583] env[62914]: DEBUG nova.virt.hardware [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.310800] env[62914]: DEBUG nova.virt.hardware [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.311114] env[62914]: DEBUG nova.virt.hardware [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.311278] env[62914]: DEBUG nova.virt.hardware [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.311431] env[62914]: DEBUG nova.virt.hardware [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.311662] env[62914]: DEBUG nova.virt.hardware [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.311793] env[62914]: DEBUG nova.virt.hardware [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.311959] env[62914]: DEBUG nova.virt.hardware [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.312144] env[62914]: DEBUG nova.virt.hardware [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.312353] env[62914]: DEBUG nova.virt.hardware [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.316017] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09ec953-f3f3-41d1-902c-a1d0753f2ce5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.323533] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43d46cf-e959-4210-bcbe-7ecc35b11672 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.382096] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f4978698-787c-4db5-b846-4d975533e897 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.569s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.383111] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a8cdf56f-2c00-40af-a806-567613f93187 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.216s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.383368] env[62914]: DEBUG nova.compute.manager [None req-a8cdf56f-2c00-40af-a806-567613f93187 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1013.384387] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8176e2e-97b8-4f3c-b62d-fb2be365b14f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.398741] env[62914]: DEBUG nova.compute.manager [None req-a8cdf56f-2c00-40af-a806-567613f93187 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62914) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1013.398741] env[62914]: DEBUG nova.objects.instance [None req-a8cdf56f-2c00-40af-a806-567613f93187 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lazy-loading 'flavor' on Instance uuid 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.459768] env[62914]: DEBUG nova.network.neutron [req-25b2e8a2-5cca-41ec-8952-69ea43e448ce req-d3affacc-a608-4245-aa3c-807fb25e46bd service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updated VIF entry in instance network info cache for port c240aa60-b896-4756-9de5-3965cfc8b21a. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1013.460207] env[62914]: DEBUG nova.network.neutron [req-25b2e8a2-5cca-41ec-8952-69ea43e448ce req-d3affacc-a608-4245-aa3c-807fb25e46bd service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updating instance_info_cache with network_info: [{"id": "c240aa60-b896-4756-9de5-3965cfc8b21a", "address": "fa:16:3e:23:5d:80", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc240aa60-b8", "ovs_interfaceid": "c240aa60-b896-4756-9de5-3965cfc8b21a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.464648] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.699s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.466888] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.205s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.467180] env[62914]: DEBUG nova.objects.instance [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lazy-loading 'resources' on Instance uuid 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.491528] env[62914]: INFO nova.scheduler.client.report [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Deleted allocations for instance 961280ed-0160-4c4e-bca4-c5f2091e9bb0 [ 1013.751303] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352936, 'name': CreateVM_Task, 'duration_secs': 0.504903} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.751501] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1013.752205] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.752375] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.752734] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1013.752934] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd684bd0-1bdc-432e-bd56-9af3cf0bb105 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.757606] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1013.757606] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ff4b5c-565f-df08-e681-e59064f0dde3" [ 1013.757606] env[62914]: _type = "Task" [ 1013.757606] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.765095] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ff4b5c-565f-df08-e681-e59064f0dde3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.823041] env[62914]: DEBUG nova.network.neutron [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Successfully updated port: f59c63fd-416a-42f8-9c9f-1b19b5003020 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1013.905972] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8cdf56f-2c00-40af-a806-567613f93187 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1013.905972] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f4a37bf-0992-4b15-83c8-0307e055de2e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.913103] env[62914]: DEBUG oslo_vmware.api [None req-a8cdf56f-2c00-40af-a806-567613f93187 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1013.913103] env[62914]: value = "task-1352937" [ 1013.913103] env[62914]: _type = "Task" [ 1013.913103] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.923716] env[62914]: DEBUG oslo_vmware.api [None req-a8cdf56f-2c00-40af-a806-567613f93187 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352937, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.963137] env[62914]: DEBUG oslo_concurrency.lockutils [req-25b2e8a2-5cca-41ec-8952-69ea43e448ce req-d3affacc-a608-4245-aa3c-807fb25e46bd service nova] Releasing lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.999424] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bc4b8a9b-32ad-4858-a431-802a4da21aa6 tempest-ServersTestJSON-212089741 tempest-ServersTestJSON-212089741-project-member] Lock "961280ed-0160-4c4e-bca4-c5f2091e9bb0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.636s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.037530] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33060300-5383-4c37-bd74-341fdce39a6c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.046050] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21aef29c-c30b-4a4b-b329-c1089543392c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.078486] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a061605-589f-41b7-b812-7881716ccf8c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.085657] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9ae4f7-ed0c-4694-b820-778f03d4eafd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.099612] env[62914]: DEBUG nova.compute.provider_tree [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.267927] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ff4b5c-565f-df08-e681-e59064f0dde3, 'name': SearchDatastore_Task, 'duration_secs': 0.012287} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.268200] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.268432] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1014.268669] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.268818] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.268998] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1014.269266] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ade4f223-adc8-4726-82ae-ba4200557f37 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.276715] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1014.276888] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1014.277560] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28d6f167-24b1-4080-bdcc-19a337463097 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.282226] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1014.282226] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ac7480-c53e-b550-5b95-21f4be24fa7c" [ 1014.282226] env[62914]: _type = "Task" [ 1014.282226] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.288907] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ac7480-c53e-b550-5b95-21f4be24fa7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.326138] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.326408] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.326451] env[62914]: DEBUG nova.network.neutron [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1014.421748] env[62914]: DEBUG oslo_vmware.api [None req-a8cdf56f-2c00-40af-a806-567613f93187 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352937, 'name': PowerOffVM_Task, 'duration_secs': 0.202001} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.422055] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8cdf56f-2c00-40af-a806-567613f93187 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1014.422246] env[62914]: DEBUG nova.compute.manager [None req-a8cdf56f-2c00-40af-a806-567613f93187 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1014.423039] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7607a83e-9592-4986-aa77-572b23185cfe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.603240] env[62914]: DEBUG nova.scheduler.client.report [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.780988] env[62914]: DEBUG nova.compute.manager [req-83ee4b9f-e8cc-4c79-9006-2368c69bad61 req-32d061b0-97ed-43b7-98b5-e32fb230daca service nova] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Received event network-vif-plugged-f59c63fd-416a-42f8-9c9f-1b19b5003020 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.781217] env[62914]: DEBUG oslo_concurrency.lockutils [req-83ee4b9f-e8cc-4c79-9006-2368c69bad61 req-32d061b0-97ed-43b7-98b5-e32fb230daca service nova] Acquiring lock "824f38d6-26c0-4922-ad3d-111b21bf08ec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.781425] env[62914]: DEBUG oslo_concurrency.lockutils [req-83ee4b9f-e8cc-4c79-9006-2368c69bad61 req-32d061b0-97ed-43b7-98b5-e32fb230daca service nova] Lock "824f38d6-26c0-4922-ad3d-111b21bf08ec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.781595] env[62914]: DEBUG oslo_concurrency.lockutils [req-83ee4b9f-e8cc-4c79-9006-2368c69bad61 req-32d061b0-97ed-43b7-98b5-e32fb230daca service nova] Lock "824f38d6-26c0-4922-ad3d-111b21bf08ec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.781761] env[62914]: DEBUG nova.compute.manager [req-83ee4b9f-e8cc-4c79-9006-2368c69bad61 req-32d061b0-97ed-43b7-98b5-e32fb230daca service nova] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] No waiting events found dispatching network-vif-plugged-f59c63fd-416a-42f8-9c9f-1b19b5003020 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.782015] env[62914]: WARNING nova.compute.manager [req-83ee4b9f-e8cc-4c79-9006-2368c69bad61 req-32d061b0-97ed-43b7-98b5-e32fb230daca service nova] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Received unexpected event network-vif-plugged-f59c63fd-416a-42f8-9c9f-1b19b5003020 for instance with vm_state building and task_state spawning. [ 1014.782205] env[62914]: DEBUG nova.compute.manager [req-83ee4b9f-e8cc-4c79-9006-2368c69bad61 req-32d061b0-97ed-43b7-98b5-e32fb230daca service nova] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Received event network-changed-f59c63fd-416a-42f8-9c9f-1b19b5003020 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.782367] env[62914]: DEBUG nova.compute.manager [req-83ee4b9f-e8cc-4c79-9006-2368c69bad61 req-32d061b0-97ed-43b7-98b5-e32fb230daca service nova] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Refreshing instance network info cache due to event network-changed-f59c63fd-416a-42f8-9c9f-1b19b5003020. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1014.782537] env[62914]: DEBUG oslo_concurrency.lockutils [req-83ee4b9f-e8cc-4c79-9006-2368c69bad61 req-32d061b0-97ed-43b7-98b5-e32fb230daca service nova] Acquiring lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.792393] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52ac7480-c53e-b550-5b95-21f4be24fa7c, 'name': SearchDatastore_Task, 'duration_secs': 0.008021} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.793151] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-037e7dae-5866-49d5-9f9f-17c6eb3b4734 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.798048] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1014.798048] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52523b5f-c57e-3b5d-5fe7-bb87c1db0633" [ 1014.798048] env[62914]: _type = "Task" [ 1014.798048] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.805724] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52523b5f-c57e-3b5d-5fe7-bb87c1db0633, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.859767] env[62914]: DEBUG nova.network.neutron [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1014.936927] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a8cdf56f-2c00-40af-a806-567613f93187 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.554s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.025130] env[62914]: DEBUG nova.network.neutron [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating instance_info_cache with network_info: [{"id": "f59c63fd-416a-42f8-9c9f-1b19b5003020", "address": "fa:16:3e:91:7a:94", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf59c63fd-41", "ovs_interfaceid": "f59c63fd-416a-42f8-9c9f-1b19b5003020", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.111528] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.644s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.134963] env[62914]: INFO nova.scheduler.client.report [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Deleted allocations for instance 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4 [ 1015.309169] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52523b5f-c57e-3b5d-5fe7-bb87c1db0633, 'name': SearchDatastore_Task, 'duration_secs': 0.009591} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.309169] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.309169] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 1028b121-ea93-4d9a-b2fd-0c2483c5c618/1028b121-ea93-4d9a-b2fd-0c2483c5c618.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1015.309336] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3df3912b-da13-48d1-80ab-8b71ba685877 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.316750] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1015.316750] env[62914]: value = "task-1352938" [ 1015.316750] env[62914]: _type = "Task" [ 1015.316750] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.325136] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352938, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.527717] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.528103] env[62914]: DEBUG nova.compute.manager [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Instance network_info: |[{"id": "f59c63fd-416a-42f8-9c9f-1b19b5003020", "address": "fa:16:3e:91:7a:94", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf59c63fd-41", "ovs_interfaceid": "f59c63fd-416a-42f8-9c9f-1b19b5003020", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1015.528437] env[62914]: DEBUG oslo_concurrency.lockutils [req-83ee4b9f-e8cc-4c79-9006-2368c69bad61 req-32d061b0-97ed-43b7-98b5-e32fb230daca service nova] Acquired lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.528622] env[62914]: DEBUG nova.network.neutron [req-83ee4b9f-e8cc-4c79-9006-2368c69bad61 req-32d061b0-97ed-43b7-98b5-e32fb230daca service nova] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Refreshing network info cache for port f59c63fd-416a-42f8-9c9f-1b19b5003020 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1015.529846] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:7a:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a8c8175-1197-4f12-baac-ef6aba95f585', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f59c63fd-416a-42f8-9c9f-1b19b5003020', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1015.538188] env[62914]: DEBUG oslo.service.loopingcall [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.541733] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1015.542374] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a5475676-c038-4516-9aaf-052845615c71 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.566023] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1015.566023] env[62914]: value = "task-1352939" [ 1015.566023] env[62914]: _type = "Task" [ 1015.566023] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.574155] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352939, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.646542] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e67773-e201-4971-9d49-6a3c4537f857 tempest-ImagesTestJSON-2078669851 tempest-ImagesTestJSON-2078669851-project-member] Lock "054f37ad-4161-4c70-aa1b-b2e69cd1e9f4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.805s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.688612] env[62914]: DEBUG nova.objects.instance [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lazy-loading 'flavor' on Instance uuid 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.826490] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352938, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.826757] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 1028b121-ea93-4d9a-b2fd-0c2483c5c618/1028b121-ea93-4d9a-b2fd-0c2483c5c618.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1015.826971] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1015.827262] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e54ef89-c705-4d03-999a-0c2a9afc542b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.833518] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1015.833518] env[62914]: value = "task-1352940" [ 1015.833518] env[62914]: _type = "Task" [ 1015.833518] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.842352] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352940, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.847806] env[62914]: DEBUG nova.network.neutron [req-83ee4b9f-e8cc-4c79-9006-2368c69bad61 req-32d061b0-97ed-43b7-98b5-e32fb230daca service nova] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updated VIF entry in instance network info cache for port f59c63fd-416a-42f8-9c9f-1b19b5003020. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1015.848237] env[62914]: DEBUG nova.network.neutron [req-83ee4b9f-e8cc-4c79-9006-2368c69bad61 req-32d061b0-97ed-43b7-98b5-e32fb230daca service nova] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating instance_info_cache with network_info: [{"id": "f59c63fd-416a-42f8-9c9f-1b19b5003020", "address": "fa:16:3e:91:7a:94", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf59c63fd-41", "ovs_interfaceid": "f59c63fd-416a-42f8-9c9f-1b19b5003020", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.075120] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352939, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.194489] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.194738] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquired lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.194971] env[62914]: DEBUG nova.network.neutron [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1016.195227] env[62914]: DEBUG nova.objects.instance [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lazy-loading 'info_cache' on Instance uuid 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.345443] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352940, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071507} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.345762] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1016.346526] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18331965-f1e7-4364-9a7c-7ab68fe93396 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.350450] env[62914]: DEBUG oslo_concurrency.lockutils [req-83ee4b9f-e8cc-4c79-9006-2368c69bad61 req-32d061b0-97ed-43b7-98b5-e32fb230daca service nova] Releasing lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.368382] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 1028b121-ea93-4d9a-b2fd-0c2483c5c618/1028b121-ea93-4d9a-b2fd-0c2483c5c618.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.368658] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-675cae3b-a8a7-4ed3-90e4-ce7d15bab0f9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.388177] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1016.388177] env[62914]: value = "task-1352941" [ 1016.388177] env[62914]: _type = "Task" [ 1016.388177] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.396577] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352941, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.576720] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352939, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.698637] env[62914]: DEBUG nova.objects.base [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Object Instance<062cd4d7-3cf1-4f31-9631-e9d1e194ad53> lazy-loaded attributes: flavor,info_cache {{(pid=62914) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1016.898300] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352941, 'name': ReconfigVM_Task, 'duration_secs': 0.267023} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.898595] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 1028b121-ea93-4d9a-b2fd-0c2483c5c618/1028b121-ea93-4d9a-b2fd-0c2483c5c618.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.899267] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-efaf9471-0cbf-4771-954b-81b8e18c4271 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.905446] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1016.905446] env[62914]: value = "task-1352942" [ 1016.905446] env[62914]: _type = "Task" [ 1016.905446] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.912891] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352942, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.077186] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352939, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.415480] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352942, 'name': Rename_Task, 'duration_secs': 0.145611} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.415782] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1017.416081] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4ebc43d-2cd6-40ce-8530-38f905ae137f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.422160] env[62914]: DEBUG nova.network.neutron [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Updating instance_info_cache with network_info: [{"id": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "address": "fa:16:3e:e8:d6:2b", "network": {"id": "74764d0b-64d3-4552-9517-ca2e87251c0b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1459372326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bb65abdffe44acdac0cdb8701e3e27b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape02166ec-c4", "ovs_interfaceid": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.424400] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1017.424400] env[62914]: value = "task-1352943" [ 1017.424400] env[62914]: _type = "Task" [ 1017.424400] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.431861] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352943, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.577879] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352939, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.926505] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Releasing lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.937631] env[62914]: DEBUG oslo_vmware.api [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352943, 'name': PowerOnVM_Task, 'duration_secs': 0.452325} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.938987] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1017.939358] env[62914]: INFO nova.compute.manager [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Took 7.45 seconds to spawn the instance on the hypervisor. [ 1017.939592] env[62914]: DEBUG nova.compute.manager [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1017.941195] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12a971b-3c9c-42a2-b0ff-b1b5bf8a3012 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.077802] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352939, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.433721] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1018.434160] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f8de285-332c-432c-83bd-48089c4c59b7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.441901] env[62914]: DEBUG oslo_vmware.api [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1018.441901] env[62914]: value = "task-1352944" [ 1018.441901] env[62914]: _type = "Task" [ 1018.441901] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.451776] env[62914]: DEBUG oslo_vmware.api [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352944, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.461884] env[62914]: INFO nova.compute.manager [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Took 12.39 seconds to build instance. [ 1018.578639] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352939, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.952014] env[62914]: DEBUG oslo_vmware.api [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352944, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.963611] env[62914]: DEBUG oslo_concurrency.lockutils [None req-26bcbb6c-faef-4367-9b89-ed6b7769df17 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.899s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.079201] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352939, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.452792] env[62914]: DEBUG oslo_vmware.api [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352944, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.579688] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352939, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.591855] env[62914]: DEBUG nova.compute.manager [req-e73b8562-5cf4-42cb-99eb-3ce1cfd9af93 req-0ddac94e-e50e-41d6-9b73-e021712cd7d2 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Received event network-changed-c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1019.592314] env[62914]: DEBUG nova.compute.manager [req-e73b8562-5cf4-42cb-99eb-3ce1cfd9af93 req-0ddac94e-e50e-41d6-9b73-e021712cd7d2 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Refreshing instance network info cache due to event network-changed-c240aa60-b896-4756-9de5-3965cfc8b21a. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1019.592631] env[62914]: DEBUG oslo_concurrency.lockutils [req-e73b8562-5cf4-42cb-99eb-3ce1cfd9af93 req-0ddac94e-e50e-41d6-9b73-e021712cd7d2 service nova] Acquiring lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.592896] env[62914]: DEBUG oslo_concurrency.lockutils [req-e73b8562-5cf4-42cb-99eb-3ce1cfd9af93 req-0ddac94e-e50e-41d6-9b73-e021712cd7d2 service nova] Acquired lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.593212] env[62914]: DEBUG nova.network.neutron [req-e73b8562-5cf4-42cb-99eb-3ce1cfd9af93 req-0ddac94e-e50e-41d6-9b73-e021712cd7d2 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Refreshing network info cache for port c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1019.954498] env[62914]: DEBUG oslo_vmware.api [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352944, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.080059] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352939, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.293790] env[62914]: DEBUG nova.network.neutron [req-e73b8562-5cf4-42cb-99eb-3ce1cfd9af93 req-0ddac94e-e50e-41d6-9b73-e021712cd7d2 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updated VIF entry in instance network info cache for port c240aa60-b896-4756-9de5-3965cfc8b21a. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1020.294187] env[62914]: DEBUG nova.network.neutron [req-e73b8562-5cf4-42cb-99eb-3ce1cfd9af93 req-0ddac94e-e50e-41d6-9b73-e021712cd7d2 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updating instance_info_cache with network_info: [{"id": "c240aa60-b896-4756-9de5-3965cfc8b21a", "address": "fa:16:3e:23:5d:80", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc240aa60-b8", "ovs_interfaceid": "c240aa60-b896-4756-9de5-3965cfc8b21a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.453907] env[62914]: DEBUG oslo_vmware.api [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352944, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.471833] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.472187] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.580570] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352939, 'name': CreateVM_Task} progress is 25%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.796702] env[62914]: DEBUG oslo_concurrency.lockutils [req-e73b8562-5cf4-42cb-99eb-3ce1cfd9af93 req-0ddac94e-e50e-41d6-9b73-e021712cd7d2 service nova] Releasing lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.954903] env[62914]: DEBUG oslo_vmware.api [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352944, 'name': PowerOnVM_Task, 'duration_secs': 2.246676} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.955203] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1020.955398] env[62914]: DEBUG nova.compute.manager [None req-4c11ad6e-3f40-4391-b6f5-2c72ade9eef6 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.956200] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a3a8c9-ca28-413a-b723-e4b26466a53d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.974879] env[62914]: DEBUG nova.compute.manager [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1021.083062] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352939, 'name': CreateVM_Task, 'duration_secs': 5.187273} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.083062] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1021.083062] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.083363] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.083624] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1021.083920] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f9c0aa7-f089-4b79-81e7-bdbca5701b3a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.088303] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1021.088303] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]520f6f60-04d1-6943-e65b-c09a0be4d5b1" [ 1021.088303] env[62914]: _type = "Task" [ 1021.088303] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.095986] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]520f6f60-04d1-6943-e65b-c09a0be4d5b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.499019] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.499019] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.499019] env[62914]: INFO nova.compute.claims [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1021.599221] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]520f6f60-04d1-6943-e65b-c09a0be4d5b1, 'name': SearchDatastore_Task, 'duration_secs': 0.01054} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.599821] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.600086] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.600332] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.600563] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.600762] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.601038] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b598aa3c-6b4a-46af-92c8-01a36c66d19a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.613622] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.613799] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1021.614509] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89123887-501d-49ad-9900-28c254964c72 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.619528] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1021.619528] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]528783cc-3db4-85b7-708b-287dfe4c1ad3" [ 1021.619528] env[62914]: _type = "Task" [ 1021.619528] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.626564] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]528783cc-3db4-85b7-708b-287dfe4c1ad3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.130223] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]528783cc-3db4-85b7-708b-287dfe4c1ad3, 'name': SearchDatastore_Task, 'duration_secs': 0.042724} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.131047] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-618b3ae2-6ad2-4958-a01d-6dcfde93744b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.136243] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1022.136243] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d30909-12cf-14d7-978e-245d6d6f78b6" [ 1022.136243] env[62914]: _type = "Task" [ 1022.136243] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.143497] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d30909-12cf-14d7-978e-245d6d6f78b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.569683] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53e31f05-eaf2-4d82-9798-37bf0936b908 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.578257] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a22cffc-68ac-48b8-8ed0-5c512a35e856 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.607629] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf90d74-a4df-4b03-b07d-dd16a03497f3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.614707] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b7d6e7-c85a-43a5-b901-fad09c8d3688 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.627313] env[62914]: DEBUG nova.compute.provider_tree [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.645622] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52d30909-12cf-14d7-978e-245d6d6f78b6, 'name': SearchDatastore_Task, 'duration_secs': 0.045712} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.645878] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.646142] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 824f38d6-26c0-4922-ad3d-111b21bf08ec/824f38d6-26c0-4922-ad3d-111b21bf08ec.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1022.646382] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-faf65333-6ac0-4b09-a678-c93ed8d8fba1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.652524] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1022.652524] env[62914]: value = "task-1352945" [ 1022.652524] env[62914]: _type = "Task" [ 1022.652524] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.659838] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352945, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.130323] env[62914]: DEBUG nova.scheduler.client.report [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.161961] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352945, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.420696} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.162235] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 824f38d6-26c0-4922-ad3d-111b21bf08ec/824f38d6-26c0-4922-ad3d-111b21bf08ec.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1023.162443] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1023.162680] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b39c25d-a4e7-4a66-bb31-348fb2257b53 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.169240] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1023.169240] env[62914]: value = "task-1352946" [ 1023.169240] env[62914]: _type = "Task" [ 1023.169240] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.176035] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352946, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.635643] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.139s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.636196] env[62914]: DEBUG nova.compute.manager [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1023.679982] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352946, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065613} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.680255] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1023.680981] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0042bc2d-ca14-4e21-8c2d-64fd8a92639c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.702080] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 824f38d6-26c0-4922-ad3d-111b21bf08ec/824f38d6-26c0-4922-ad3d-111b21bf08ec.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.702332] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aac2af7b-2175-4bdc-99af-d7f81805cba6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.720680] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1023.720680] env[62914]: value = "task-1352947" [ 1023.720680] env[62914]: _type = "Task" [ 1023.720680] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.728102] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352947, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.141906] env[62914]: DEBUG nova.compute.utils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1024.143358] env[62914]: DEBUG nova.compute.manager [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1024.143744] env[62914]: DEBUG nova.network.neutron [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1024.179262] env[62914]: DEBUG nova.policy [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80089586cfda4cd2bdfc24d9f556239c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09d51090ac5d41fc966c7e132f288824', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 1024.229895] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352947, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.424604] env[62914]: DEBUG nova.network.neutron [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Successfully created port: 1481dffb-497c-4588-8bb1-12ce0445c7c0 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1024.649831] env[62914]: DEBUG nova.compute.manager [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1024.730729] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352947, 'name': ReconfigVM_Task, 'duration_secs': 0.9339} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.731354] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 824f38d6-26c0-4922-ad3d-111b21bf08ec/824f38d6-26c0-4922-ad3d-111b21bf08ec.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.732189] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be09d29b-2a47-4a49-8278-7e62e5856536 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.738136] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1024.738136] env[62914]: value = "task-1352948" [ 1024.738136] env[62914]: _type = "Task" [ 1024.738136] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.745981] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352948, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.249074] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352948, 'name': Rename_Task, 'duration_secs': 0.137049} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.249074] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1025.249074] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1eb60597-e89e-43bc-ab79-1b867baa686f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.254901] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1025.254901] env[62914]: value = "task-1352949" [ 1025.254901] env[62914]: _type = "Task" [ 1025.254901] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.264769] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352949, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.657767] env[62914]: DEBUG nova.compute.manager [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1025.685242] env[62914]: DEBUG nova.virt.hardware [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1025.685641] env[62914]: DEBUG nova.virt.hardware [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1025.685901] env[62914]: DEBUG nova.virt.hardware [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1025.686258] env[62914]: DEBUG nova.virt.hardware [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1025.686511] env[62914]: DEBUG nova.virt.hardware [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1025.686804] env[62914]: DEBUG nova.virt.hardware [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1025.687189] env[62914]: DEBUG nova.virt.hardware [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1025.687486] env[62914]: DEBUG nova.virt.hardware [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1025.687791] env[62914]: DEBUG nova.virt.hardware [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1025.688080] env[62914]: DEBUG nova.virt.hardware [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1025.688375] env[62914]: DEBUG nova.virt.hardware [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1025.689672] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361058fc-b802-48c5-929c-f5d11f54d3f0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.701375] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f809ce31-917d-495e-8adc-0553edcac5e4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.764715] env[62914]: DEBUG oslo_vmware.api [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352949, 'name': PowerOnVM_Task, 'duration_secs': 0.478611} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.764984] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1025.765216] env[62914]: INFO nova.compute.manager [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Took 12.49 seconds to spawn the instance on the hypervisor. [ 1025.765402] env[62914]: DEBUG nova.compute.manager [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1025.766195] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e78ee8-e685-4075-8bce-946c1ce2ea47 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.008511] env[62914]: DEBUG nova.compute.manager [req-d94c4630-c9aa-4a23-b266-9e017665f508 req-2ebba568-397f-4629-b0f7-3d4dc167830d service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Received event network-vif-plugged-1481dffb-497c-4588-8bb1-12ce0445c7c0 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.008511] env[62914]: DEBUG oslo_concurrency.lockutils [req-d94c4630-c9aa-4a23-b266-9e017665f508 req-2ebba568-397f-4629-b0f7-3d4dc167830d service nova] Acquiring lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.008511] env[62914]: DEBUG oslo_concurrency.lockutils [req-d94c4630-c9aa-4a23-b266-9e017665f508 req-2ebba568-397f-4629-b0f7-3d4dc167830d service nova] Lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.008626] env[62914]: DEBUG oslo_concurrency.lockutils [req-d94c4630-c9aa-4a23-b266-9e017665f508 req-2ebba568-397f-4629-b0f7-3d4dc167830d service nova] Lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.008751] env[62914]: DEBUG nova.compute.manager [req-d94c4630-c9aa-4a23-b266-9e017665f508 req-2ebba568-397f-4629-b0f7-3d4dc167830d service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] No waiting events found dispatching network-vif-plugged-1481dffb-497c-4588-8bb1-12ce0445c7c0 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1026.008919] env[62914]: WARNING nova.compute.manager [req-d94c4630-c9aa-4a23-b266-9e017665f508 req-2ebba568-397f-4629-b0f7-3d4dc167830d service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Received unexpected event network-vif-plugged-1481dffb-497c-4588-8bb1-12ce0445c7c0 for instance with vm_state building and task_state spawning. [ 1026.098311] env[62914]: DEBUG nova.network.neutron [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Successfully updated port: 1481dffb-497c-4588-8bb1-12ce0445c7c0 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1026.282199] env[62914]: INFO nova.compute.manager [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Took 17.66 seconds to build instance. [ 1026.600230] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.600476] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.600599] env[62914]: DEBUG nova.network.neutron [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1026.784935] env[62914]: DEBUG oslo_concurrency.lockutils [None req-db967545-7290-4c9a-a99f-ee3088e88aa1 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "824f38d6-26c0-4922-ad3d-111b21bf08ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.168s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.136739] env[62914]: DEBUG nova.network.neutron [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1027.268234] env[62914]: DEBUG nova.network.neutron [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updating instance_info_cache with network_info: [{"id": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "address": "fa:16:3e:63:cc:75", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1481dffb-49", "ovs_interfaceid": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.386253] env[62914]: DEBUG nova.compute.manager [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Stashing vm_state: active {{(pid=62914) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1027.771034] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.771298] env[62914]: DEBUG nova.compute.manager [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Instance network_info: |[{"id": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "address": "fa:16:3e:63:cc:75", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1481dffb-49", "ovs_interfaceid": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1027.771793] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:cc:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e31264e2-3e0a-4dfb-ba1f-6389d7d47548', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1481dffb-497c-4588-8bb1-12ce0445c7c0', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1027.779243] env[62914]: DEBUG oslo.service.loopingcall [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.779464] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1027.779690] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e8d1777d-0509-404c-a0b5-6b9d2edc3244 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.799957] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1027.799957] env[62914]: value = "task-1352950" [ 1027.799957] env[62914]: _type = "Task" [ 1027.799957] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.807905] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352950, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.908041] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.908174] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.031437] env[62914]: DEBUG nova.compute.manager [req-7084ac51-30e6-476a-ad28-6c2bec177a49 req-1d382802-fc51-4aed-81fa-56c57eb43790 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Received event network-changed-1481dffb-497c-4588-8bb1-12ce0445c7c0 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.031684] env[62914]: DEBUG nova.compute.manager [req-7084ac51-30e6-476a-ad28-6c2bec177a49 req-1d382802-fc51-4aed-81fa-56c57eb43790 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Refreshing instance network info cache due to event network-changed-1481dffb-497c-4588-8bb1-12ce0445c7c0. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1028.031907] env[62914]: DEBUG oslo_concurrency.lockutils [req-7084ac51-30e6-476a-ad28-6c2bec177a49 req-1d382802-fc51-4aed-81fa-56c57eb43790 service nova] Acquiring lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.032182] env[62914]: DEBUG oslo_concurrency.lockutils [req-7084ac51-30e6-476a-ad28-6c2bec177a49 req-1d382802-fc51-4aed-81fa-56c57eb43790 service nova] Acquired lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.032433] env[62914]: DEBUG nova.network.neutron [req-7084ac51-30e6-476a-ad28-6c2bec177a49 req-1d382802-fc51-4aed-81fa-56c57eb43790 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Refreshing network info cache for port 1481dffb-497c-4588-8bb1-12ce0445c7c0 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1028.309355] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352950, 'name': CreateVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.415024] env[62914]: INFO nova.compute.claims [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1028.762963] env[62914]: DEBUG nova.network.neutron [req-7084ac51-30e6-476a-ad28-6c2bec177a49 req-1d382802-fc51-4aed-81fa-56c57eb43790 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updated VIF entry in instance network info cache for port 1481dffb-497c-4588-8bb1-12ce0445c7c0. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1028.763383] env[62914]: DEBUG nova.network.neutron [req-7084ac51-30e6-476a-ad28-6c2bec177a49 req-1d382802-fc51-4aed-81fa-56c57eb43790 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updating instance_info_cache with network_info: [{"id": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "address": "fa:16:3e:63:cc:75", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1481dffb-49", "ovs_interfaceid": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.809535] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352950, 'name': CreateVM_Task, 'duration_secs': 0.851802} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.809840] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1028.810373] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.810541] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.810898] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1028.811169] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3026cca-ee84-45a3-8aea-65c86ddf0ba3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.815396] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1028.815396] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a877d6-0a1b-b68f-16b3-50cd68b7070a" [ 1028.815396] env[62914]: _type = "Task" [ 1028.815396] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.822457] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a877d6-0a1b-b68f-16b3-50cd68b7070a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.921377] env[62914]: INFO nova.compute.resource_tracker [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating resource usage from migration 963c9fd3-1c2e-486d-91c0-c6596348d652 [ 1028.989054] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9afe6d85-84bd-4568-99e8-ce3e20eb7103 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.996775] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780cae31-5015-4695-a623-447ce691cb89 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.025316] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-787575cb-bf06-44a6-9be3-e90fe823a83e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.032369] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be85375-0821-40e9-b81c-c75b4618de5b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.046354] env[62914]: DEBUG nova.compute.provider_tree [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.265738] env[62914]: DEBUG oslo_concurrency.lockutils [req-7084ac51-30e6-476a-ad28-6c2bec177a49 req-1d382802-fc51-4aed-81fa-56c57eb43790 service nova] Releasing lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.325182] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a877d6-0a1b-b68f-16b3-50cd68b7070a, 'name': SearchDatastore_Task, 'duration_secs': 0.010291} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.325472] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.325716] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.325979] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.326147] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.326329] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1029.326574] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a694e525-7b93-42a9-82f4-d014c41a22fd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.334098] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1029.334271] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1029.334904] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d986ea35-82da-4683-b8ff-8690f9d4468a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.339426] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1029.339426] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]526e4420-84fc-d8f0-ca22-5eaaa024f51a" [ 1029.339426] env[62914]: _type = "Task" [ 1029.339426] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.346334] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]526e4420-84fc-d8f0-ca22-5eaaa024f51a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.550011] env[62914]: DEBUG nova.scheduler.client.report [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.849506] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]526e4420-84fc-d8f0-ca22-5eaaa024f51a, 'name': SearchDatastore_Task, 'duration_secs': 0.00779} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.850318] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-327bc17a-3213-4369-a32e-56d6492076a8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.855317] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1029.855317] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a523f6-67ed-88e8-5dba-85188839ef01" [ 1029.855317] env[62914]: _type = "Task" [ 1029.855317] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.862219] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a523f6-67ed-88e8-5dba-85188839ef01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.054714] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.146s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.054909] env[62914]: INFO nova.compute.manager [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Migrating [ 1030.365472] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a523f6-67ed-88e8-5dba-85188839ef01, 'name': SearchDatastore_Task, 'duration_secs': 0.008722} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.365691] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.365988] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 58c2cc4a-153d-437b-8822-d3e389bd1a7f/58c2cc4a-153d-437b-8822-d3e389bd1a7f.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1030.366269] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5f683c3-26b4-43bc-a232-5988d4e30aec {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.372666] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1030.372666] env[62914]: value = "task-1352951" [ 1030.372666] env[62914]: _type = "Task" [ 1030.372666] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.379657] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352951, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.569966] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.570266] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.570451] env[62914]: DEBUG nova.network.neutron [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1030.883030] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352951, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.449317} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.883311] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 58c2cc4a-153d-437b-8822-d3e389bd1a7f/58c2cc4a-153d-437b-8822-d3e389bd1a7f.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1030.883375] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1030.883686] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea7a98e1-956e-44a3-b3e5-dff269fc086f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.889676] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1030.889676] env[62914]: value = "task-1352952" [ 1030.889676] env[62914]: _type = "Task" [ 1030.889676] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.897872] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352952, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.248798] env[62914]: DEBUG nova.network.neutron [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating instance_info_cache with network_info: [{"id": "f59c63fd-416a-42f8-9c9f-1b19b5003020", "address": "fa:16:3e:91:7a:94", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf59c63fd-41", "ovs_interfaceid": "f59c63fd-416a-42f8-9c9f-1b19b5003020", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.399391] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352952, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063789} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.399616] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1031.400377] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5fad8a-ddf6-4700-9d9c-0e40bab9ddf6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.421566] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 58c2cc4a-153d-437b-8822-d3e389bd1a7f/58c2cc4a-153d-437b-8822-d3e389bd1a7f.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.421842] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39459ba6-9ea3-4020-b91d-85bba522f626 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.440955] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1031.440955] env[62914]: value = "task-1352953" [ 1031.440955] env[62914]: _type = "Task" [ 1031.440955] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.449609] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352953, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.751419] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.951253] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352953, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.229355] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.229548] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.229697] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Starting heal instance info cache {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1032.451971] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352953, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.953226] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352953, 'name': ReconfigVM_Task, 'duration_secs': 1.139678} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.953584] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 58c2cc4a-153d-437b-8822-d3e389bd1a7f/58c2cc4a-153d-437b-8822-d3e389bd1a7f.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.954249] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7076fe6-8f5c-4e20-993f-129596b98314 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.960542] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1032.960542] env[62914]: value = "task-1352954" [ 1032.960542] env[62914]: _type = "Task" [ 1032.960542] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.968064] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352954, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.265191] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b9026fb-7e38-4314-b4bd-7b772e59c5df {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.285722] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating instance '824f38d6-26c0-4922-ad3d-111b21bf08ec' progress to 0 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1033.470603] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352954, 'name': Rename_Task, 'duration_secs': 0.166408} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.470918] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1033.472468] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de7c4b0f-3eb0-4ab5-96c6-5b4bd43ae131 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.478123] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1033.478123] env[62914]: value = "task-1352955" [ 1033.478123] env[62914]: _type = "Task" [ 1033.478123] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.485314] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352955, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.792502] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1033.792796] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c89f75eb-1b32-43cb-9dc1-279e65731a2c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.801225] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1033.801225] env[62914]: value = "task-1352956" [ 1033.801225] env[62914]: _type = "Task" [ 1033.801225] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.809352] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352956, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.987926] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352955, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.264887] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.265057] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquired lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.265205] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Forcefully refreshing network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1034.311513] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352956, 'name': PowerOffVM_Task, 'duration_secs': 0.15349} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.311757] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1034.311947] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating instance '824f38d6-26c0-4922-ad3d-111b21bf08ec' progress to 17 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1034.489624] env[62914]: DEBUG oslo_vmware.api [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352955, 'name': PowerOnVM_Task, 'duration_secs': 0.521394} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.489900] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1034.490111] env[62914]: INFO nova.compute.manager [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Took 8.83 seconds to spawn the instance on the hypervisor. [ 1034.490295] env[62914]: DEBUG nova.compute.manager [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.491155] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd6949e-ffae-4b76-a2be-7b2dbdebcbba {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.818097] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1034.818341] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1034.818508] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1034.818696] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1034.818844] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1034.820040] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1034.820040] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1034.820040] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1034.820040] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1034.820040] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1034.820040] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1034.825128] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e324f73a-1b71-438e-af1d-02db8eefa6bd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.840699] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1034.840699] env[62914]: value = "task-1352957" [ 1034.840699] env[62914]: _type = "Task" [ 1034.840699] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.848694] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352957, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.010393] env[62914]: INFO nova.compute.manager [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Took 13.53 seconds to build instance. [ 1035.350109] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352957, 'name': ReconfigVM_Task, 'duration_secs': 0.146578} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.350590] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating instance '824f38d6-26c0-4922-ad3d-111b21bf08ec' progress to 33 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1035.512899] env[62914]: DEBUG oslo_concurrency.lockutils [None req-633cc751-e47f-4cd4-95ea-6ac3ee297e8d tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.041s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.514159] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Updating instance_info_cache with network_info: [{"id": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "address": "fa:16:3e:e8:d6:2b", "network": {"id": "74764d0b-64d3-4552-9517-ca2e87251c0b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1459372326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bb65abdffe44acdac0cdb8701e3e27b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape02166ec-c4", "ovs_interfaceid": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.857265] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1035.857522] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1035.857700] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1035.857891] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1035.858115] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1035.858300] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1035.858514] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1035.858681] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1035.858895] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1035.859088] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1035.859274] env[62914]: DEBUG nova.virt.hardware [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1035.865183] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Reconfiguring VM instance instance-00000065 to detach disk 2000 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1035.865475] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b2b92c5-4982-4fb8-bd0c-8b10dce994c2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.884690] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1035.884690] env[62914]: value = "task-1352958" [ 1035.884690] env[62914]: _type = "Task" [ 1035.884690] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.893118] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352958, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.016943] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Releasing lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.017239] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Updated the network info_cache for instance {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1036.017436] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1036.017611] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1036.017758] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1036.017959] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1036.018069] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1036.018258] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1036.018384] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62914) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1036.018555] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager.update_available_resource {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1036.028135] env[62914]: DEBUG nova.compute.manager [req-3608fd5f-ae6e-4a6e-b448-28bc67263d48 req-1096eaf8-d2ec-405c-b688-99947732e0ed service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Received event network-changed-c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1036.028330] env[62914]: DEBUG nova.compute.manager [req-3608fd5f-ae6e-4a6e-b448-28bc67263d48 req-1096eaf8-d2ec-405c-b688-99947732e0ed service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Refreshing instance network info cache due to event network-changed-c240aa60-b896-4756-9de5-3965cfc8b21a. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1036.028551] env[62914]: DEBUG oslo_concurrency.lockutils [req-3608fd5f-ae6e-4a6e-b448-28bc67263d48 req-1096eaf8-d2ec-405c-b688-99947732e0ed service nova] Acquiring lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.028699] env[62914]: DEBUG oslo_concurrency.lockutils [req-3608fd5f-ae6e-4a6e-b448-28bc67263d48 req-1096eaf8-d2ec-405c-b688-99947732e0ed service nova] Acquired lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.028864] env[62914]: DEBUG nova.network.neutron [req-3608fd5f-ae6e-4a6e-b448-28bc67263d48 req-1096eaf8-d2ec-405c-b688-99947732e0ed service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Refreshing network info cache for port c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1036.395500] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352958, 'name': ReconfigVM_Task, 'duration_secs': 0.159169} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.396119] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Reconfigured VM instance instance-00000065 to detach disk 2000 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1036.396762] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f1fd1d-5365-4dac-b9fb-2eb717f39a46 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.418412] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 824f38d6-26c0-4922-ad3d-111b21bf08ec/824f38d6-26c0-4922-ad3d-111b21bf08ec.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1036.418833] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52799784-d8a2-4aeb-89e0-c366e473e38b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.436838] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1036.436838] env[62914]: value = "task-1352959" [ 1036.436838] env[62914]: _type = "Task" [ 1036.436838] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.447251] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352959, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.522252] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.522901] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.522901] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.522901] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62914) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1036.523910] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82fe91cf-367b-4d3b-9bbe-0075c0828f02 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.533883] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427da2b5-19f0-4be1-b62e-8bd47e7d075e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.549777] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a45443d-43ad-47e7-8dce-567ddc27b8af {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.557017] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c271b2e6-ea77-4075-93eb-fd1113de9760 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.586338] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181158MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62914) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1036.586502] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.586719] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.773824] env[62914]: DEBUG nova.network.neutron [req-3608fd5f-ae6e-4a6e-b448-28bc67263d48 req-1096eaf8-d2ec-405c-b688-99947732e0ed service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updated VIF entry in instance network info cache for port c240aa60-b896-4756-9de5-3965cfc8b21a. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1036.774217] env[62914]: DEBUG nova.network.neutron [req-3608fd5f-ae6e-4a6e-b448-28bc67263d48 req-1096eaf8-d2ec-405c-b688-99947732e0ed service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updating instance_info_cache with network_info: [{"id": "c240aa60-b896-4756-9de5-3965cfc8b21a", "address": "fa:16:3e:23:5d:80", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc240aa60-b8", "ovs_interfaceid": "c240aa60-b896-4756-9de5-3965cfc8b21a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.947262] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352959, 'name': ReconfigVM_Task, 'duration_secs': 0.301287} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.947585] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 824f38d6-26c0-4922-ad3d-111b21bf08ec/824f38d6-26c0-4922-ad3d-111b21bf08ec.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1036.947877] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating instance '824f38d6-26c0-4922-ad3d-111b21bf08ec' progress to 50 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1037.277238] env[62914]: DEBUG oslo_concurrency.lockutils [req-3608fd5f-ae6e-4a6e-b448-28bc67263d48 req-1096eaf8-d2ec-405c-b688-99947732e0ed service nova] Releasing lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.454585] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b35b7a-4e22-4fe9-870e-ba2b6c0ec4e5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.475032] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0001ab65-4272-40b5-a765-ad2e119c7e43 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.492563] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating instance '824f38d6-26c0-4922-ad3d-111b21bf08ec' progress to 67 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1037.597041] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Applying migration context for instance 824f38d6-26c0-4922-ad3d-111b21bf08ec as it has an incoming, in-progress migration 963c9fd3-1c2e-486d-91c0-c6596348d652. Migration status is migrating {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1037.597840] env[62914]: INFO nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating resource usage from migration 963c9fd3-1c2e-486d-91c0-c6596348d652 [ 1037.686261] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1037.686424] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 1028b121-ea93-4d9a-b2fd-0c2483c5c618 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1037.686546] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 58c2cc4a-153d-437b-8822-d3e389bd1a7f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1037.686662] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Migration 963c9fd3-1c2e-486d-91c0-c6596348d652 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1037.686776] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 824f38d6-26c0-4922-ad3d-111b21bf08ec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1037.686968] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1037.687121] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1536MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1037.750644] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3bc44d-539a-4fed-a192-49809d872801 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.758007] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecbdb54a-48ae-4751-8f3f-8f9aa118779d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.786553] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bab6fa4-c272-4f1a-b735-6bf0f16fd2ee {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.793496] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c0969e-9054-4838-9e6d-50acbe53c3e8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.805909] env[62914]: DEBUG nova.compute.provider_tree [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.030402] env[62914]: DEBUG nova.network.neutron [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Port f59c63fd-416a-42f8-9c9f-1b19b5003020 binding to destination host cpu-1 is already ACTIVE {{(pid=62914) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1038.060729] env[62914]: DEBUG nova.compute.manager [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Received event network-changed-1481dffb-497c-4588-8bb1-12ce0445c7c0 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1038.060922] env[62914]: DEBUG nova.compute.manager [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Refreshing instance network info cache due to event network-changed-1481dffb-497c-4588-8bb1-12ce0445c7c0. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1038.061175] env[62914]: DEBUG oslo_concurrency.lockutils [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] Acquiring lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.061330] env[62914]: DEBUG oslo_concurrency.lockutils [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] Acquired lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.061496] env[62914]: DEBUG nova.network.neutron [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Refreshing network info cache for port 1481dffb-497c-4588-8bb1-12ce0445c7c0 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1038.308585] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.760714] env[62914]: DEBUG nova.network.neutron [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updated VIF entry in instance network info cache for port 1481dffb-497c-4588-8bb1-12ce0445c7c0. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1038.761128] env[62914]: DEBUG nova.network.neutron [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updating instance_info_cache with network_info: [{"id": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "address": "fa:16:3e:63:cc:75", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1481dffb-49", "ovs_interfaceid": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.814870] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1038.815028] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.228s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.052572] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "824f38d6-26c0-4922-ad3d-111b21bf08ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.052741] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "824f38d6-26c0-4922-ad3d-111b21bf08ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.052782] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "824f38d6-26c0-4922-ad3d-111b21bf08ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.263659] env[62914]: DEBUG oslo_concurrency.lockutils [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] Releasing lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.263938] env[62914]: DEBUG nova.compute.manager [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Received event network-changed-1481dffb-497c-4588-8bb1-12ce0445c7c0 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.264126] env[62914]: DEBUG nova.compute.manager [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Refreshing instance network info cache due to event network-changed-1481dffb-497c-4588-8bb1-12ce0445c7c0. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1039.264341] env[62914]: DEBUG oslo_concurrency.lockutils [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] Acquiring lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.264565] env[62914]: DEBUG oslo_concurrency.lockutils [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] Acquired lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.264754] env[62914]: DEBUG nova.network.neutron [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Refreshing network info cache for port 1481dffb-497c-4588-8bb1-12ce0445c7c0 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1039.941694] env[62914]: DEBUG nova.network.neutron [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updated VIF entry in instance network info cache for port 1481dffb-497c-4588-8bb1-12ce0445c7c0. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1039.942072] env[62914]: DEBUG nova.network.neutron [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updating instance_info_cache with network_info: [{"id": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "address": "fa:16:3e:63:cc:75", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1481dffb-49", "ovs_interfaceid": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.086356] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.086551] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.086731] env[62914]: DEBUG nova.network.neutron [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1040.444733] env[62914]: DEBUG oslo_concurrency.lockutils [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] Releasing lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.445021] env[62914]: DEBUG nova.compute.manager [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Received event network-changed-c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.445202] env[62914]: DEBUG nova.compute.manager [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Refreshing instance network info cache due to event network-changed-c240aa60-b896-4756-9de5-3965cfc8b21a. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1040.445418] env[62914]: DEBUG oslo_concurrency.lockutils [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] Acquiring lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.445564] env[62914]: DEBUG oslo_concurrency.lockutils [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] Acquired lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.445727] env[62914]: DEBUG nova.network.neutron [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Refreshing network info cache for port c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1040.805590] env[62914]: DEBUG nova.network.neutron [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating instance_info_cache with network_info: [{"id": "f59c63fd-416a-42f8-9c9f-1b19b5003020", "address": "fa:16:3e:91:7a:94", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf59c63fd-41", "ovs_interfaceid": "f59c63fd-416a-42f8-9c9f-1b19b5003020", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.145319] env[62914]: DEBUG nova.network.neutron [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updated VIF entry in instance network info cache for port c240aa60-b896-4756-9de5-3965cfc8b21a. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1041.145697] env[62914]: DEBUG nova.network.neutron [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updating instance_info_cache with network_info: [{"id": "c240aa60-b896-4756-9de5-3965cfc8b21a", "address": "fa:16:3e:23:5d:80", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc240aa60-b8", "ovs_interfaceid": "c240aa60-b896-4756-9de5-3965cfc8b21a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.308881] env[62914]: DEBUG oslo_concurrency.lockutils [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.648683] env[62914]: DEBUG oslo_concurrency.lockutils [req-2b9d8b32-0585-4e3b-b095-228353dd35ec req-3a1602de-b0c3-49f8-85d1-623879a8c6f2 service nova] Releasing lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.832506] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed292f03-3ee6-442d-9845-3c08370e70f2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.851480] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0722948f-2f4f-4323-a685-970f9c5865f0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.858432] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating instance '824f38d6-26c0-4922-ad3d-111b21bf08ec' progress to 83 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1042.365290] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1042.365713] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-020ac651-62bd-4018-8885-07363938dd80 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.373491] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1042.373491] env[62914]: value = "task-1352960" [ 1042.373491] env[62914]: _type = "Task" [ 1042.373491] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.381385] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352960, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.882580] env[62914]: DEBUG oslo_vmware.api [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352960, 'name': PowerOnVM_Task, 'duration_secs': 0.45757} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.882812] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1042.882994] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-80a8803e-7d74-4f7e-b536-09adbdaed893 tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating instance '824f38d6-26c0-4922-ad3d-111b21bf08ec' progress to 100 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1044.667697] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "824f38d6-26c0-4922-ad3d-111b21bf08ec" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.668058] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "824f38d6-26c0-4922-ad3d-111b21bf08ec" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.668199] env[62914]: DEBUG nova.compute.manager [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Going to confirm migration 2 {{(pid=62914) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1045.223034] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.223253] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquired lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.223471] env[62914]: DEBUG nova.network.neutron [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1045.223685] env[62914]: DEBUG nova.objects.instance [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lazy-loading 'info_cache' on Instance uuid 824f38d6-26c0-4922-ad3d-111b21bf08ec {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.408331] env[62914]: DEBUG nova.network.neutron [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating instance_info_cache with network_info: [{"id": "f59c63fd-416a-42f8-9c9f-1b19b5003020", "address": "fa:16:3e:91:7a:94", "network": {"id": "d19106e2-893b-4009-a249-0421b506c044", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-466904719-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dcda44f0d6a455bb17966e7c5061190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a8c8175-1197-4f12-baac-ef6aba95f585", "external-id": "nsx-vlan-transportzone-832", "segmentation_id": 832, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf59c63fd-41", "ovs_interfaceid": "f59c63fd-416a-42f8-9c9f-1b19b5003020", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.910792] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Releasing lock "refresh_cache-824f38d6-26c0-4922-ad3d-111b21bf08ec" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.911021] env[62914]: DEBUG nova.objects.instance [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lazy-loading 'migration_context' on Instance uuid 824f38d6-26c0-4922-ad3d-111b21bf08ec {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.414214] env[62914]: DEBUG nova.objects.base [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Object Instance<824f38d6-26c0-4922-ad3d-111b21bf08ec> lazy-loaded attributes: info_cache,migration_context {{(pid=62914) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1047.415566] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c6b0003-f8bf-4ca6-bbb8-56d5a88ba4f2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.434808] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de90e64e-fa3f-4ecb-9a88-dc539537e39f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.439861] env[62914]: DEBUG oslo_vmware.api [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1047.439861] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]524d6cf8-6856-b8fd-5b2b-5e0741adb13f" [ 1047.439861] env[62914]: _type = "Task" [ 1047.439861] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.447013] env[62914]: DEBUG oslo_vmware.api [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]524d6cf8-6856-b8fd-5b2b-5e0741adb13f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.950514] env[62914]: DEBUG oslo_vmware.api [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]524d6cf8-6856-b8fd-5b2b-5e0741adb13f, 'name': SearchDatastore_Task, 'duration_secs': 0.008087} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.950809] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.951068] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.530884] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6257c01b-d3a1-4ba8-8aa2-b02c564b1845 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.539073] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2982bed-24d1-42f5-92c1-bbe248f7bb1b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.570308] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0720239b-3e6b-406b-a4ef-e83f6e73b204 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.578355] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb7c615-be93-441d-a860-6d545f4da53a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.592371] env[62914]: DEBUG nova.compute.provider_tree [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.096307] env[62914]: DEBUG nova.scheduler.client.report [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.564582] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "interface-1028b121-ea93-4d9a-b2fd-0c2483c5c618-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.565197] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-1028b121-ea93-4d9a-b2fd-0c2483c5c618-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.565291] env[62914]: DEBUG nova.objects.instance [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'flavor' on Instance uuid 1028b121-ea93-4d9a-b2fd-0c2483c5c618 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.106311] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.155s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.167632] env[62914]: DEBUG nova.objects.instance [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'pci_requests' on Instance uuid 1028b121-ea93-4d9a-b2fd-0c2483c5c618 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.659730] env[62914]: INFO nova.scheduler.client.report [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleted allocation for migration 963c9fd3-1c2e-486d-91c0-c6596348d652 [ 1050.669735] env[62914]: DEBUG nova.objects.base [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Object Instance<1028b121-ea93-4d9a-b2fd-0c2483c5c618> lazy-loaded attributes: flavor,pci_requests {{(pid=62914) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1050.669989] env[62914]: DEBUG nova.network.neutron [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1050.733315] env[62914]: DEBUG nova.policy [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80089586cfda4cd2bdfc24d9f556239c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09d51090ac5d41fc966c7e132f288824', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 1051.165695] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "824f38d6-26c0-4922-ad3d-111b21bf08ec" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.497s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.098770] env[62914]: DEBUG nova.compute.manager [req-70f06342-0630-47d2-a520-f0cc02cafaa0 req-9ebace1e-d3e6-4046-a175-6c550ea038db service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Received event network-vif-plugged-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1052.099019] env[62914]: DEBUG oslo_concurrency.lockutils [req-70f06342-0630-47d2-a520-f0cc02cafaa0 req-9ebace1e-d3e6-4046-a175-6c550ea038db service nova] Acquiring lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.099229] env[62914]: DEBUG oslo_concurrency.lockutils [req-70f06342-0630-47d2-a520-f0cc02cafaa0 req-9ebace1e-d3e6-4046-a175-6c550ea038db service nova] Lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.099412] env[62914]: DEBUG oslo_concurrency.lockutils [req-70f06342-0630-47d2-a520-f0cc02cafaa0 req-9ebace1e-d3e6-4046-a175-6c550ea038db service nova] Lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.099552] env[62914]: DEBUG nova.compute.manager [req-70f06342-0630-47d2-a520-f0cc02cafaa0 req-9ebace1e-d3e6-4046-a175-6c550ea038db service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] No waiting events found dispatching network-vif-plugged-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1052.099725] env[62914]: WARNING nova.compute.manager [req-70f06342-0630-47d2-a520-f0cc02cafaa0 req-9ebace1e-d3e6-4046-a175-6c550ea038db service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Received unexpected event network-vif-plugged-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64 for instance with vm_state active and task_state None. [ 1052.177329] env[62914]: DEBUG nova.network.neutron [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Successfully updated port: b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1052.192342] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "824f38d6-26c0-4922-ad3d-111b21bf08ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.192609] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "824f38d6-26c0-4922-ad3d-111b21bf08ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.192822] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "824f38d6-26c0-4922-ad3d-111b21bf08ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.193017] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "824f38d6-26c0-4922-ad3d-111b21bf08ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.193274] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "824f38d6-26c0-4922-ad3d-111b21bf08ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.195185] env[62914]: INFO nova.compute.manager [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Terminating instance [ 1052.196978] env[62914]: DEBUG nova.compute.manager [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1052.197198] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1052.198030] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5225637-76a2-4ae2-a70d-a9a75cbf8214 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.206091] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1052.206325] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09d8d0e0-73f9-41b7-b0cf-edcd42eaece2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.212228] env[62914]: DEBUG oslo_vmware.api [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1052.212228] env[62914]: value = "task-1352961" [ 1052.212228] env[62914]: _type = "Task" [ 1052.212228] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.219439] env[62914]: DEBUG oslo_vmware.api [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352961, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.680058] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.680058] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.680373] env[62914]: DEBUG nova.network.neutron [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1052.721770] env[62914]: DEBUG oslo_vmware.api [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352961, 'name': PowerOffVM_Task, 'duration_secs': 0.180655} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.722024] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1052.722199] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1052.722450] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-69eed9e7-d7a1-49e9-adab-90d9776c153f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.781215] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1052.781437] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1052.781626] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleting the datastore file [datastore1] 824f38d6-26c0-4922-ad3d-111b21bf08ec {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1052.781880] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-917aa692-043b-4266-9347-839c12b0ae78 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.788336] env[62914]: DEBUG oslo_vmware.api [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for the task: (returnval){ [ 1052.788336] env[62914]: value = "task-1352963" [ 1052.788336] env[62914]: _type = "Task" [ 1052.788336] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.795600] env[62914]: DEBUG oslo_vmware.api [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352963, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.213497] env[62914]: WARNING nova.network.neutron [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] f59e1bed-468c-492e-a501-34925e946a66 already exists in list: networks containing: ['f59e1bed-468c-492e-a501-34925e946a66']. ignoring it [ 1053.299902] env[62914]: DEBUG oslo_vmware.api [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Task: {'id': task-1352963, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138116} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.300122] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1053.300319] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1053.300499] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1053.300676] env[62914]: INFO nova.compute.manager [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1053.300936] env[62914]: DEBUG oslo.service.loopingcall [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.301146] env[62914]: DEBUG nova.compute.manager [-] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1053.301242] env[62914]: DEBUG nova.network.neutron [-] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1053.484292] env[62914]: DEBUG nova.network.neutron [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updating instance_info_cache with network_info: [{"id": "c240aa60-b896-4756-9de5-3965cfc8b21a", "address": "fa:16:3e:23:5d:80", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc240aa60-b8", "ovs_interfaceid": "c240aa60-b896-4756-9de5-3965cfc8b21a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64", "address": "fa:16:3e:e2:d4:fc", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb575b8e5-b8", "ovs_interfaceid": "b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.566867] env[62914]: DEBUG nova.compute.manager [req-343205c9-c0ae-4231-b682-f08d8e349e94 req-19061e68-9e99-49de-b285-6a07287b23c8 service nova] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Received event network-vif-deleted-f59c63fd-416a-42f8-9c9f-1b19b5003020 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.567099] env[62914]: INFO nova.compute.manager [req-343205c9-c0ae-4231-b682-f08d8e349e94 req-19061e68-9e99-49de-b285-6a07287b23c8 service nova] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Neutron deleted interface f59c63fd-416a-42f8-9c9f-1b19b5003020; detaching it from the instance and deleting it from the info cache [ 1053.567263] env[62914]: DEBUG nova.network.neutron [req-343205c9-c0ae-4231-b682-f08d8e349e94 req-19061e68-9e99-49de-b285-6a07287b23c8 service nova] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.987137] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.987809] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.987970] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.988868] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4648f6ae-fb82-4c2a-95f8-7a3ea9e3654b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.005565] env[62914]: DEBUG nova.virt.hardware [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1054.005790] env[62914]: DEBUG nova.virt.hardware [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1054.005947] env[62914]: DEBUG nova.virt.hardware [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1054.006245] env[62914]: DEBUG nova.virt.hardware [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1054.006417] env[62914]: DEBUG nova.virt.hardware [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.006572] env[62914]: DEBUG nova.virt.hardware [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1054.006786] env[62914]: DEBUG nova.virt.hardware [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1054.006948] env[62914]: DEBUG nova.virt.hardware [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1054.007134] env[62914]: DEBUG nova.virt.hardware [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1054.007661] env[62914]: DEBUG nova.virt.hardware [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1054.007661] env[62914]: DEBUG nova.virt.hardware [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1054.013653] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Reconfiguring VM to attach interface {{(pid=62914) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1054.013942] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2b130da-0955-4dcf-9abd-57e946333b1c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.030879] env[62914]: DEBUG oslo_vmware.api [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1054.030879] env[62914]: value = "task-1352964" [ 1054.030879] env[62914]: _type = "Task" [ 1054.030879] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.038406] env[62914]: DEBUG oslo_vmware.api [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352964, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.043756] env[62914]: DEBUG nova.network.neutron [-] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.069959] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-40c85ec9-2759-4c8b-9e90-441e5a265bfd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.078749] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ebf614-3b37-4767-bef6-cf6e1d4f653e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.102696] env[62914]: DEBUG nova.compute.manager [req-343205c9-c0ae-4231-b682-f08d8e349e94 req-19061e68-9e99-49de-b285-6a07287b23c8 service nova] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Detach interface failed, port_id=f59c63fd-416a-42f8-9c9f-1b19b5003020, reason: Instance 824f38d6-26c0-4922-ad3d-111b21bf08ec could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1054.124154] env[62914]: DEBUG nova.compute.manager [req-239ae991-619b-45e6-9861-9c668bded5c6 req-d7c659cd-6f0b-461d-b27a-ee8ac732e93c service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Received event network-changed-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1054.124356] env[62914]: DEBUG nova.compute.manager [req-239ae991-619b-45e6-9861-9c668bded5c6 req-d7c659cd-6f0b-461d-b27a-ee8ac732e93c service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Refreshing instance network info cache due to event network-changed-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1054.124583] env[62914]: DEBUG oslo_concurrency.lockutils [req-239ae991-619b-45e6-9861-9c668bded5c6 req-d7c659cd-6f0b-461d-b27a-ee8ac732e93c service nova] Acquiring lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.124733] env[62914]: DEBUG oslo_concurrency.lockutils [req-239ae991-619b-45e6-9861-9c668bded5c6 req-d7c659cd-6f0b-461d-b27a-ee8ac732e93c service nova] Acquired lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.124893] env[62914]: DEBUG nova.network.neutron [req-239ae991-619b-45e6-9861-9c668bded5c6 req-d7c659cd-6f0b-461d-b27a-ee8ac732e93c service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Refreshing network info cache for port b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1054.541825] env[62914]: DEBUG oslo_vmware.api [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352964, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.546366] env[62914]: INFO nova.compute.manager [-] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Took 1.25 seconds to deallocate network for instance. [ 1054.819506] env[62914]: DEBUG nova.network.neutron [req-239ae991-619b-45e6-9861-9c668bded5c6 req-d7c659cd-6f0b-461d-b27a-ee8ac732e93c service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updated VIF entry in instance network info cache for port b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1054.819937] env[62914]: DEBUG nova.network.neutron [req-239ae991-619b-45e6-9861-9c668bded5c6 req-d7c659cd-6f0b-461d-b27a-ee8ac732e93c service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updating instance_info_cache with network_info: [{"id": "c240aa60-b896-4756-9de5-3965cfc8b21a", "address": "fa:16:3e:23:5d:80", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc240aa60-b8", "ovs_interfaceid": "c240aa60-b896-4756-9de5-3965cfc8b21a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64", "address": "fa:16:3e:e2:d4:fc", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb575b8e5-b8", "ovs_interfaceid": "b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.040721] env[62914]: DEBUG oslo_vmware.api [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352964, 'name': ReconfigVM_Task, 'duration_secs': 0.519308} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.041256] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.041560] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Reconfigured VM to attach interface {{(pid=62914) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1055.054040] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.054040] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.054163] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.076776] env[62914]: INFO nova.scheduler.client.report [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Deleted allocations for instance 824f38d6-26c0-4922-ad3d-111b21bf08ec [ 1055.322956] env[62914]: DEBUG oslo_concurrency.lockutils [req-239ae991-619b-45e6-9861-9c668bded5c6 req-d7c659cd-6f0b-461d-b27a-ee8ac732e93c service nova] Releasing lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.546571] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7438a3a2-970e-4871-af1c-98bc122f9c72 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-1028b121-ea93-4d9a-b2fd-0c2483c5c618-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 5.981s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.584498] env[62914]: DEBUG oslo_concurrency.lockutils [None req-02bb30cc-b82e-4ade-a1bd-eb409585076f tempest-DeleteServersTestJSON-1396104689 tempest-DeleteServersTestJSON-1396104689-project-member] Lock "824f38d6-26c0-4922-ad3d-111b21bf08ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.392s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.487123] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.487411] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.754721] env[62914]: DEBUG oslo_concurrency.lockutils [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "interface-1028b121-ea93-4d9a-b2fd-0c2483c5c618-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.755114] env[62914]: DEBUG oslo_concurrency.lockutils [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-1028b121-ea93-4d9a-b2fd-0c2483c5c618-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.990632] env[62914]: INFO nova.compute.manager [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Detaching volume 96b3ee28-51d2-440e-a63d-387302c4807d [ 1057.020068] env[62914]: INFO nova.virt.block_device [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Attempting to driver detach volume 96b3ee28-51d2-440e-a63d-387302c4807d from mountpoint /dev/sdb [ 1057.020387] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Volume detach. Driver type: vmdk {{(pid=62914) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1057.020591] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288265', 'volume_id': '96b3ee28-51d2-440e-a63d-387302c4807d', 'name': 'volume-96b3ee28-51d2-440e-a63d-387302c4807d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '062cd4d7-3cf1-4f31-9631-e9d1e194ad53', 'attached_at': '', 'detached_at': '', 'volume_id': '96b3ee28-51d2-440e-a63d-387302c4807d', 'serial': '96b3ee28-51d2-440e-a63d-387302c4807d'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1057.021759] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f83e337-cd1e-41b9-9f0e-8ea430b4bad3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.042553] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27bc32df-a426-4b0d-a5cb-1767acd40711 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.049294] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d009af15-1f7d-44ef-b3bb-868b52eaab15 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.070222] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceda9faf-5b3f-4ed3-b1f8-ec99183470eb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.086115] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] The volume has not been displaced from its original location: [datastore1] volume-96b3ee28-51d2-440e-a63d-387302c4807d/volume-96b3ee28-51d2-440e-a63d-387302c4807d.vmdk. No consolidation needed. {{(pid=62914) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1057.091184] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Reconfiguring VM instance instance-0000005d to detach disk 2001 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1057.091453] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cdc049b9-1282-4c12-8082-2e73ca913fce {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.108857] env[62914]: DEBUG oslo_vmware.api [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1057.108857] env[62914]: value = "task-1352966" [ 1057.108857] env[62914]: _type = "Task" [ 1057.108857] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.116105] env[62914]: DEBUG oslo_vmware.api [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352966, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.258096] env[62914]: DEBUG oslo_concurrency.lockutils [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.258332] env[62914]: DEBUG oslo_concurrency.lockutils [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.259228] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c363319-bd02-4fc0-a391-ace95e668ba8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.276517] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e32dfcb-7226-4b1b-b647-2f86bb5a5f81 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.302095] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Reconfiguring VM to detach interface {{(pid=62914) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1057.302397] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c38eadea-314a-4461-adea-51c24b5cac6d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.320960] env[62914]: DEBUG oslo_vmware.api [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1057.320960] env[62914]: value = "task-1352967" [ 1057.320960] env[62914]: _type = "Task" [ 1057.320960] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.331647] env[62914]: DEBUG oslo_vmware.api [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352967, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.619335] env[62914]: DEBUG oslo_vmware.api [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352966, 'name': ReconfigVM_Task, 'duration_secs': 0.259445} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.619562] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Reconfigured VM instance instance-0000005d to detach disk 2001 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1057.624121] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5cc5278-db32-4194-8c64-816c0c24879a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.638228] env[62914]: DEBUG oslo_vmware.api [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1057.638228] env[62914]: value = "task-1352968" [ 1057.638228] env[62914]: _type = "Task" [ 1057.638228] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.645459] env[62914]: DEBUG oslo_vmware.api [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352968, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.830100] env[62914]: DEBUG oslo_vmware.api [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.150163] env[62914]: DEBUG oslo_vmware.api [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352968, 'name': ReconfigVM_Task, 'duration_secs': 0.127453} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.150595] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288265', 'volume_id': '96b3ee28-51d2-440e-a63d-387302c4807d', 'name': 'volume-96b3ee28-51d2-440e-a63d-387302c4807d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '062cd4d7-3cf1-4f31-9631-e9d1e194ad53', 'attached_at': '', 'detached_at': '', 'volume_id': '96b3ee28-51d2-440e-a63d-387302c4807d', 'serial': '96b3ee28-51d2-440e-a63d-387302c4807d'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1058.331685] env[62914]: DEBUG oslo_vmware.api [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.691872] env[62914]: DEBUG nova.objects.instance [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lazy-loading 'flavor' on Instance uuid 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.833364] env[62914]: DEBUG oslo_vmware.api [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.332904] env[62914]: DEBUG oslo_vmware.api [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.655828] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0fe47461-39f0-421a-bd27-01fbec10643d tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.699039] env[62914]: DEBUG oslo_concurrency.lockutils [None req-5f3f5be2-d4d0-4d43-888e-440674879299 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.211s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.700225] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0fe47461-39f0-421a-bd27-01fbec10643d tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.045s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.700427] env[62914]: DEBUG nova.compute.manager [None req-0fe47461-39f0-421a-bd27-01fbec10643d tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1059.701513] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be7ba34-e808-48b2-b0fe-cdf77b04b562 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.708952] env[62914]: DEBUG nova.compute.manager [None req-0fe47461-39f0-421a-bd27-01fbec10643d tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62914) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1059.709564] env[62914]: DEBUG nova.objects.instance [None req-0fe47461-39f0-421a-bd27-01fbec10643d tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lazy-loading 'flavor' on Instance uuid 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.833906] env[62914]: DEBUG oslo_vmware.api [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.214123] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fe47461-39f0-421a-bd27-01fbec10643d tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1060.214415] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff2be636-581d-4c21-b268-b39c50b1e76c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.222367] env[62914]: DEBUG oslo_vmware.api [None req-0fe47461-39f0-421a-bd27-01fbec10643d tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1060.222367] env[62914]: value = "task-1352969" [ 1060.222367] env[62914]: _type = "Task" [ 1060.222367] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.230508] env[62914]: DEBUG oslo_vmware.api [None req-0fe47461-39f0-421a-bd27-01fbec10643d tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352969, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.334762] env[62914]: DEBUG oslo_vmware.api [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.731957] env[62914]: DEBUG oslo_vmware.api [None req-0fe47461-39f0-421a-bd27-01fbec10643d tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352969, 'name': PowerOffVM_Task, 'duration_secs': 0.165302} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.732248] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fe47461-39f0-421a-bd27-01fbec10643d tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1060.732426] env[62914]: DEBUG nova.compute.manager [None req-0fe47461-39f0-421a-bd27-01fbec10643d tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1060.733548] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b123ebc-ccc4-44f0-8a93-c920010e6538 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.835840] env[62914]: DEBUG oslo_vmware.api [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.243890] env[62914]: DEBUG oslo_concurrency.lockutils [None req-0fe47461-39f0-421a-bd27-01fbec10643d tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.335803] env[62914]: DEBUG oslo_vmware.api [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.836575] env[62914]: DEBUG oslo_vmware.api [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.126020] env[62914]: DEBUG nova.objects.instance [None req-33884511-d14a-451d-8bdc-f8f3e1405796 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lazy-loading 'flavor' on Instance uuid 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1062.336848] env[62914]: DEBUG oslo_vmware.api [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.630744] env[62914]: DEBUG oslo_concurrency.lockutils [None req-33884511-d14a-451d-8bdc-f8f3e1405796 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.630918] env[62914]: DEBUG oslo_concurrency.lockutils [None req-33884511-d14a-451d-8bdc-f8f3e1405796 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquired lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.631115] env[62914]: DEBUG nova.network.neutron [None req-33884511-d14a-451d-8bdc-f8f3e1405796 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1062.631314] env[62914]: DEBUG nova.objects.instance [None req-33884511-d14a-451d-8bdc-f8f3e1405796 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lazy-loading 'info_cache' on Instance uuid 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1062.837143] env[62914]: DEBUG oslo_vmware.api [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352967, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.135049] env[62914]: DEBUG nova.objects.base [None req-33884511-d14a-451d-8bdc-f8f3e1405796 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Object Instance<062cd4d7-3cf1-4f31-9631-e9d1e194ad53> lazy-loaded attributes: flavor,info_cache {{(pid=62914) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1063.337476] env[62914]: DEBUG oslo_vmware.api [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352967, 'name': ReconfigVM_Task, 'duration_secs': 5.772768} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.337731] env[62914]: DEBUG oslo_concurrency.lockutils [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.337948] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Reconfigured VM to detach interface {{(pid=62914) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1063.859240] env[62914]: DEBUG nova.network.neutron [None req-33884511-d14a-451d-8bdc-f8f3e1405796 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Updating instance_info_cache with network_info: [{"id": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "address": "fa:16:3e:e8:d6:2b", "network": {"id": "74764d0b-64d3-4552-9517-ca2e87251c0b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1459372326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bb65abdffe44acdac0cdb8701e3e27b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape02166ec-c4", "ovs_interfaceid": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.362508] env[62914]: DEBUG nova.compute.manager [req-0b613567-4ab2-4288-9dce-4372d5c04a53 req-7023d189-a0a3-4170-b59b-bb1858a70d4b service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Received event network-changed-c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1064.363193] env[62914]: DEBUG nova.compute.manager [req-0b613567-4ab2-4288-9dce-4372d5c04a53 req-7023d189-a0a3-4170-b59b-bb1858a70d4b service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Refreshing instance network info cache due to event network-changed-c240aa60-b896-4756-9de5-3965cfc8b21a. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1064.363553] env[62914]: DEBUG oslo_concurrency.lockutils [req-0b613567-4ab2-4288-9dce-4372d5c04a53 req-7023d189-a0a3-4170-b59b-bb1858a70d4b service nova] Acquiring lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.363746] env[62914]: DEBUG oslo_concurrency.lockutils [req-0b613567-4ab2-4288-9dce-4372d5c04a53 req-7023d189-a0a3-4170-b59b-bb1858a70d4b service nova] Acquired lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.363946] env[62914]: DEBUG nova.network.neutron [req-0b613567-4ab2-4288-9dce-4372d5c04a53 req-7023d189-a0a3-4170-b59b-bb1858a70d4b service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Refreshing network info cache for port c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1064.365784] env[62914]: DEBUG oslo_concurrency.lockutils [None req-33884511-d14a-451d-8bdc-f8f3e1405796 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Releasing lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.609630] env[62914]: DEBUG oslo_concurrency.lockutils [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.870658] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-33884511-d14a-451d-8bdc-f8f3e1405796 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1064.870989] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7a493804-2597-4fe6-920f-33f302adc4ab {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.878991] env[62914]: DEBUG oslo_vmware.api [None req-33884511-d14a-451d-8bdc-f8f3e1405796 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1064.878991] env[62914]: value = "task-1352970" [ 1064.878991] env[62914]: _type = "Task" [ 1064.878991] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.886601] env[62914]: DEBUG oslo_vmware.api [None req-33884511-d14a-451d-8bdc-f8f3e1405796 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352970, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.034338] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "interface-58c2cc4a-153d-437b-8822-d3e389bd1a7f-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.034611] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-58c2cc4a-153d-437b-8822-d3e389bd1a7f-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.034983] env[62914]: DEBUG nova.objects.instance [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'flavor' on Instance uuid 58c2cc4a-153d-437b-8822-d3e389bd1a7f {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.081387] env[62914]: DEBUG nova.network.neutron [req-0b613567-4ab2-4288-9dce-4372d5c04a53 req-7023d189-a0a3-4170-b59b-bb1858a70d4b service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updated VIF entry in instance network info cache for port c240aa60-b896-4756-9de5-3965cfc8b21a. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1065.081788] env[62914]: DEBUG nova.network.neutron [req-0b613567-4ab2-4288-9dce-4372d5c04a53 req-7023d189-a0a3-4170-b59b-bb1858a70d4b service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updating instance_info_cache with network_info: [{"id": "c240aa60-b896-4756-9de5-3965cfc8b21a", "address": "fa:16:3e:23:5d:80", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc240aa60-b8", "ovs_interfaceid": "c240aa60-b896-4756-9de5-3965cfc8b21a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64", "address": "fa:16:3e:e2:d4:fc", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb575b8e5-b8", "ovs_interfaceid": "b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.388476] env[62914]: DEBUG oslo_vmware.api [None req-33884511-d14a-451d-8bdc-f8f3e1405796 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352970, 'name': PowerOnVM_Task, 'duration_secs': 0.380531} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.388818] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-33884511-d14a-451d-8bdc-f8f3e1405796 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.389026] env[62914]: DEBUG nova.compute.manager [None req-33884511-d14a-451d-8bdc-f8f3e1405796 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.389872] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78cc86c-6f6d-4d91-9d35-f93ce7067496 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.584047] env[62914]: DEBUG oslo_concurrency.lockutils [req-0b613567-4ab2-4288-9dce-4372d5c04a53 req-7023d189-a0a3-4170-b59b-bb1858a70d4b service nova] Releasing lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.584580] env[62914]: DEBUG oslo_concurrency.lockutils [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.584778] env[62914]: DEBUG nova.network.neutron [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1065.628226] env[62914]: DEBUG nova.objects.instance [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'pci_requests' on Instance uuid 58c2cc4a-153d-437b-8822-d3e389bd1a7f {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.130912] env[62914]: DEBUG nova.objects.base [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Object Instance<58c2cc4a-153d-437b-8822-d3e389bd1a7f> lazy-loaded attributes: flavor,pci_requests {{(pid=62914) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1066.131221] env[62914]: DEBUG nova.network.neutron [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1066.199989] env[62914]: DEBUG nova.policy [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80089586cfda4cd2bdfc24d9f556239c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09d51090ac5d41fc966c7e132f288824', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 1066.270904] env[62914]: INFO nova.network.neutron [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Port b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1066.271440] env[62914]: DEBUG nova.network.neutron [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updating instance_info_cache with network_info: [{"id": "c240aa60-b896-4756-9de5-3965cfc8b21a", "address": "fa:16:3e:23:5d:80", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc240aa60-b8", "ovs_interfaceid": "c240aa60-b896-4756-9de5-3965cfc8b21a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.390123] env[62914]: DEBUG nova.compute.manager [req-9e818f6a-fb83-4486-b1f3-0f08b8869a99 req-8981350e-b001-46c6-9cdb-17fa62ac7479 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Received event network-changed-1481dffb-497c-4588-8bb1-12ce0445c7c0 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1066.390123] env[62914]: DEBUG nova.compute.manager [req-9e818f6a-fb83-4486-b1f3-0f08b8869a99 req-8981350e-b001-46c6-9cdb-17fa62ac7479 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Refreshing instance network info cache due to event network-changed-1481dffb-497c-4588-8bb1-12ce0445c7c0. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1066.390123] env[62914]: DEBUG oslo_concurrency.lockutils [req-9e818f6a-fb83-4486-b1f3-0f08b8869a99 req-8981350e-b001-46c6-9cdb-17fa62ac7479 service nova] Acquiring lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.390123] env[62914]: DEBUG oslo_concurrency.lockutils [req-9e818f6a-fb83-4486-b1f3-0f08b8869a99 req-8981350e-b001-46c6-9cdb-17fa62ac7479 service nova] Acquired lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.390283] env[62914]: DEBUG nova.network.neutron [req-9e818f6a-fb83-4486-b1f3-0f08b8869a99 req-8981350e-b001-46c6-9cdb-17fa62ac7479 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Refreshing network info cache for port 1481dffb-497c-4588-8bb1-12ce0445c7c0 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1066.774458] env[62914]: DEBUG oslo_concurrency.lockutils [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "refresh_cache-1028b121-ea93-4d9a-b2fd-0c2483c5c618" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.086734] env[62914]: DEBUG nova.network.neutron [req-9e818f6a-fb83-4486-b1f3-0f08b8869a99 req-8981350e-b001-46c6-9cdb-17fa62ac7479 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updated VIF entry in instance network info cache for port 1481dffb-497c-4588-8bb1-12ce0445c7c0. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1067.087118] env[62914]: DEBUG nova.network.neutron [req-9e818f6a-fb83-4486-b1f3-0f08b8869a99 req-8981350e-b001-46c6-9cdb-17fa62ac7479 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updating instance_info_cache with network_info: [{"id": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "address": "fa:16:3e:63:cc:75", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1481dffb-49", "ovs_interfaceid": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.279058] env[62914]: DEBUG oslo_concurrency.lockutils [None req-638c83fa-324a-4079-a722-4e6beed5a6bb tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-1028b121-ea93-4d9a-b2fd-0c2483c5c618-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.524s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.589954] env[62914]: DEBUG oslo_concurrency.lockutils [req-9e818f6a-fb83-4486-b1f3-0f08b8869a99 req-8981350e-b001-46c6-9cdb-17fa62ac7479 service nova] Releasing lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.672595] env[62914]: DEBUG nova.network.neutron [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Successfully updated port: b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1068.176014] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.176267] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.176466] env[62914]: DEBUG nova.network.neutron [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1068.417059] env[62914]: DEBUG nova.compute.manager [req-c963532b-f431-4ad8-a46c-9a3f8026ab3f req-7f37c1a5-d5c2-41aa-9740-b98e0e6438a4 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Received event network-vif-plugged-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.417310] env[62914]: DEBUG oslo_concurrency.lockutils [req-c963532b-f431-4ad8-a46c-9a3f8026ab3f req-7f37c1a5-d5c2-41aa-9740-b98e0e6438a4 service nova] Acquiring lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.417522] env[62914]: DEBUG oslo_concurrency.lockutils [req-c963532b-f431-4ad8-a46c-9a3f8026ab3f req-7f37c1a5-d5c2-41aa-9740-b98e0e6438a4 service nova] Lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.417696] env[62914]: DEBUG oslo_concurrency.lockutils [req-c963532b-f431-4ad8-a46c-9a3f8026ab3f req-7f37c1a5-d5c2-41aa-9740-b98e0e6438a4 service nova] Lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.417866] env[62914]: DEBUG nova.compute.manager [req-c963532b-f431-4ad8-a46c-9a3f8026ab3f req-7f37c1a5-d5c2-41aa-9740-b98e0e6438a4 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] No waiting events found dispatching network-vif-plugged-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1068.418158] env[62914]: WARNING nova.compute.manager [req-c963532b-f431-4ad8-a46c-9a3f8026ab3f req-7f37c1a5-d5c2-41aa-9740-b98e0e6438a4 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Received unexpected event network-vif-plugged-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64 for instance with vm_state active and task_state None. [ 1068.418392] env[62914]: DEBUG nova.compute.manager [req-c963532b-f431-4ad8-a46c-9a3f8026ab3f req-7f37c1a5-d5c2-41aa-9740-b98e0e6438a4 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Received event network-changed-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.418596] env[62914]: DEBUG nova.compute.manager [req-c963532b-f431-4ad8-a46c-9a3f8026ab3f req-7f37c1a5-d5c2-41aa-9740-b98e0e6438a4 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Refreshing instance network info cache due to event network-changed-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1068.418754] env[62914]: DEBUG oslo_concurrency.lockutils [req-c963532b-f431-4ad8-a46c-9a3f8026ab3f req-7f37c1a5-d5c2-41aa-9740-b98e0e6438a4 service nova] Acquiring lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.718017] env[62914]: WARNING nova.network.neutron [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] f59e1bed-468c-492e-a501-34925e946a66 already exists in list: networks containing: ['f59e1bed-468c-492e-a501-34925e946a66']. ignoring it [ 1068.972940] env[62914]: DEBUG nova.network.neutron [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updating instance_info_cache with network_info: [{"id": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "address": "fa:16:3e:63:cc:75", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1481dffb-49", "ovs_interfaceid": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64", "address": "fa:16:3e:e2:d4:fc", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb575b8e5-b8", "ovs_interfaceid": "b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.475944] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.476575] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.476741] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.477032] env[62914]: DEBUG oslo_concurrency.lockutils [req-c963532b-f431-4ad8-a46c-9a3f8026ab3f req-7f37c1a5-d5c2-41aa-9740-b98e0e6438a4 service nova] Acquired lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.477222] env[62914]: DEBUG nova.network.neutron [req-c963532b-f431-4ad8-a46c-9a3f8026ab3f req-7f37c1a5-d5c2-41aa-9740-b98e0e6438a4 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Refreshing network info cache for port b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1069.478908] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dae86c3-7229-492e-a6c2-39e13a43f1c2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.499499] env[62914]: DEBUG nova.virt.hardware [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1069.499722] env[62914]: DEBUG nova.virt.hardware [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1069.499879] env[62914]: DEBUG nova.virt.hardware [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1069.500079] env[62914]: DEBUG nova.virt.hardware [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1069.500236] env[62914]: DEBUG nova.virt.hardware [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1069.500385] env[62914]: DEBUG nova.virt.hardware [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1069.500593] env[62914]: DEBUG nova.virt.hardware [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1069.500754] env[62914]: DEBUG nova.virt.hardware [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1069.500922] env[62914]: DEBUG nova.virt.hardware [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1069.501098] env[62914]: DEBUG nova.virt.hardware [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1069.501282] env[62914]: DEBUG nova.virt.hardware [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1069.507484] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Reconfiguring VM to attach interface {{(pid=62914) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1069.507965] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4595e6f7-0ba8-4f63-a1bb-578097328070 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.525672] env[62914]: DEBUG oslo_vmware.api [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1069.525672] env[62914]: value = "task-1352971" [ 1069.525672] env[62914]: _type = "Task" [ 1069.525672] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.533995] env[62914]: DEBUG oslo_vmware.api [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352971, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.035752] env[62914]: DEBUG oslo_vmware.api [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352971, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.375026] env[62914]: DEBUG nova.network.neutron [req-c963532b-f431-4ad8-a46c-9a3f8026ab3f req-7f37c1a5-d5c2-41aa-9740-b98e0e6438a4 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updated VIF entry in instance network info cache for port b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1070.375490] env[62914]: DEBUG nova.network.neutron [req-c963532b-f431-4ad8-a46c-9a3f8026ab3f req-7f37c1a5-d5c2-41aa-9740-b98e0e6438a4 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updating instance_info_cache with network_info: [{"id": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "address": "fa:16:3e:63:cc:75", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1481dffb-49", "ovs_interfaceid": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64", "address": "fa:16:3e:e2:d4:fc", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb575b8e5-b8", "ovs_interfaceid": "b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.535832] env[62914]: DEBUG oslo_vmware.api [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352971, 'name': ReconfigVM_Task, 'duration_secs': 0.76658} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.536306] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.536549] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Reconfigured VM to attach interface {{(pid=62914) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1070.878102] env[62914]: DEBUG oslo_concurrency.lockutils [req-c963532b-f431-4ad8-a46c-9a3f8026ab3f req-7f37c1a5-d5c2-41aa-9740-b98e0e6438a4 service nova] Releasing lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.041864] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4b91b64e-7152-410c-b846-8cdcebc20842 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-58c2cc4a-153d-437b-8822-d3e389bd1a7f-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.007s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.901997] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.902393] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.902544] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.902706] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Cleaning up deleted instances {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1072.415011] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] There are 39 instances to clean {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1072.415279] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 824f38d6-26c0-4922-ad3d-111b21bf08ec] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.735376] env[62914]: DEBUG oslo_concurrency.lockutils [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "interface-58c2cc4a-153d-437b-8822-d3e389bd1a7f-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.735659] env[62914]: DEBUG oslo_concurrency.lockutils [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-58c2cc4a-153d-437b-8822-d3e389bd1a7f-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.918581] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 961280ed-0160-4c4e-bca4-c5f2091e9bb0] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.238736] env[62914]: DEBUG oslo_concurrency.lockutils [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.239368] env[62914]: DEBUG oslo_concurrency.lockutils [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.240271] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f293a4-503d-4e54-9ab6-3fb04dfb9e96 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.257596] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cbff88f-034a-4eef-b0f3-0c2dd4cefd8e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.283513] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Reconfiguring VM to detach interface {{(pid=62914) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1073.283754] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ba1b38d-c948-4e45-8193-4ba584a2cf79 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.303205] env[62914]: DEBUG oslo_vmware.api [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1073.303205] env[62914]: value = "task-1352972" [ 1073.303205] env[62914]: _type = "Task" [ 1073.303205] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.310658] env[62914]: DEBUG oslo_vmware.api [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352972, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.421675] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 054f37ad-4161-4c70-aa1b-b2e69cd1e9f4] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.814911] env[62914]: DEBUG oslo_vmware.api [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352972, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.924837] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 84a7822d-d46d-4147-ad4c-b9402aa072f4] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.314280] env[62914]: DEBUG oslo_vmware.api [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352972, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.428187] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 6cbbf7a7-a4fb-4c61-bf8b-a1300c66c509] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.814743] env[62914]: DEBUG oslo_vmware.api [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352972, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.931468] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 5e8f452f-68e6-4fe6-9955-20359d90851b] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.318308] env[62914]: DEBUG oslo_vmware.api [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352972, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.435374] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: a2a13ec9-30f3-4c04-b10b-f9a9c01534bd] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.815909] env[62914]: DEBUG oslo_vmware.api [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352972, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.938753] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 10c348bd-d85d-49b8-9b0f-216142f664d1] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.316812] env[62914]: DEBUG oslo_vmware.api [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352972, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.441982] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 2575bc0d-cb47-4dab-a219-f84254bda47d] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.817573] env[62914]: DEBUG oslo_vmware.api [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352972, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.945542] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 412c706a-31ce-4a33-914d-5762b02b4c63] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.318185] env[62914]: DEBUG oslo_vmware.api [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352972, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.449044] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 4bec4c88-bd1e-4cdf-9c8a-eee56088479b] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.820544] env[62914]: DEBUG oslo_vmware.api [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352972, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.952425] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: f616eac4-7c75-484e-bb64-5e227a08df8c] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.320425] env[62914]: DEBUG oslo_vmware.api [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352972, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.456276] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 1e527707-758c-4006-af83-9c739b9645ed] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.820926] env[62914]: DEBUG oslo_vmware.api [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352972, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.960807] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 1f7b1609-08e0-4439-aea0-b4a2ec808780] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.320968] env[62914]: DEBUG oslo_vmware.api [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352972, 'name': ReconfigVM_Task, 'duration_secs': 5.724758} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.321220] env[62914]: DEBUG oslo_concurrency.lockutils [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.321404] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Reconfigured VM to detach interface {{(pid=62914) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1079.464418] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: b8aebca4-2724-419e-aaf7-d180e299a459] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.968272] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: ab036e1f-a62a-490a-818c-2e1256c781f6] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.472137] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: f14f4ab4-5ef1-4f2c-ae97-ac924a446df0] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.623562] env[62914]: DEBUG oslo_concurrency.lockutils [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.623766] env[62914]: DEBUG oslo_concurrency.lockutils [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquired lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.623946] env[62914]: DEBUG nova.network.neutron [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1080.975064] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: ef27919a-f205-49d9-88d8-b350fcf5cfac] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.101350] env[62914]: DEBUG oslo_concurrency.lockutils [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.101564] env[62914]: DEBUG oslo_concurrency.lockutils [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.101772] env[62914]: DEBUG oslo_concurrency.lockutils [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.101958] env[62914]: DEBUG oslo_concurrency.lockutils [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.102149] env[62914]: DEBUG oslo_concurrency.lockutils [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.105224] env[62914]: INFO nova.compute.manager [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Terminating instance [ 1081.106941] env[62914]: DEBUG nova.compute.manager [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1081.107157] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1081.107979] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32733f88-28aa-4ea7-bfab-99ce397c1523 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.115601] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1081.115818] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc9a31d7-fdb3-4b25-8b33-e12125818716 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.121954] env[62914]: DEBUG oslo_vmware.api [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1081.121954] env[62914]: value = "task-1352973" [ 1081.121954] env[62914]: _type = "Task" [ 1081.121954] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.131726] env[62914]: DEBUG oslo_vmware.api [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352973, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.334706] env[62914]: INFO nova.network.neutron [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Port b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1081.335101] env[62914]: DEBUG nova.network.neutron [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updating instance_info_cache with network_info: [{"id": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "address": "fa:16:3e:63:cc:75", "network": {"id": "f59e1bed-468c-492e-a501-34925e946a66", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534938473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09d51090ac5d41fc966c7e132f288824", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31264e2-3e0a-4dfb-ba1f-6389d7d47548", "external-id": "nsx-vlan-transportzone-233", "segmentation_id": 233, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1481dffb-49", "ovs_interfaceid": "1481dffb-497c-4588-8bb1-12ce0445c7c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.478669] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: aec39e18-7796-4be3-af74-478df3a78f8f] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.631922] env[62914]: DEBUG oslo_vmware.api [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352973, 'name': PowerOffVM_Task, 'duration_secs': 0.190234} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.632202] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1081.632376] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1081.632622] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3324cf57-87a2-425b-b011-872cbb0f4dc2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.695912] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1081.696146] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1081.696335] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Deleting the datastore file [datastore2] 58c2cc4a-153d-437b-8822-d3e389bd1a7f {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1081.696593] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb97553c-dad8-40ee-afc6-d604b4693d1f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.702321] env[62914]: DEBUG oslo_vmware.api [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1081.702321] env[62914]: value = "task-1352975" [ 1081.702321] env[62914]: _type = "Task" [ 1081.702321] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.709868] env[62914]: DEBUG oslo_vmware.api [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352975, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.838419] env[62914]: DEBUG oslo_concurrency.lockutils [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Releasing lock "refresh_cache-58c2cc4a-153d-437b-8822-d3e389bd1a7f" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.981380] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: c9b94435-6546-464a-b3ba-c9e685157059] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.211743] env[62914]: DEBUG oslo_vmware.api [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352975, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160874} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.212016] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1082.212231] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1082.212417] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1082.212600] env[62914]: INFO nova.compute.manager [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1082.212843] env[62914]: DEBUG oslo.service.loopingcall [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1082.213056] env[62914]: DEBUG nova.compute.manager [-] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1082.213156] env[62914]: DEBUG nova.network.neutron [-] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1082.342097] env[62914]: DEBUG oslo_concurrency.lockutils [None req-157b306c-228a-40e2-80ad-ad78afc97978 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "interface-58c2cc4a-153d-437b-8822-d3e389bd1a7f-b575b8e5-b8a9-42a1-8fc1-7edd2c7f8b64" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.606s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.484646] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 47e6f142-bdb2-4397-9b94-d962e68ee6bb] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.851337] env[62914]: DEBUG nova.compute.manager [req-d14621bc-b48e-40eb-be5b-47e16a5d0f15 req-68394a30-92b7-4cfb-8b8d-e1e91d0000b0 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Received event network-vif-deleted-1481dffb-497c-4588-8bb1-12ce0445c7c0 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1082.851572] env[62914]: INFO nova.compute.manager [req-d14621bc-b48e-40eb-be5b-47e16a5d0f15 req-68394a30-92b7-4cfb-8b8d-e1e91d0000b0 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Neutron deleted interface 1481dffb-497c-4588-8bb1-12ce0445c7c0; detaching it from the instance and deleting it from the info cache [ 1082.851766] env[62914]: DEBUG nova.network.neutron [req-d14621bc-b48e-40eb-be5b-47e16a5d0f15 req-68394a30-92b7-4cfb-8b8d-e1e91d0000b0 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.987405] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 22de53a6-096b-4e8e-af91-f78588ea4c60] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.329606] env[62914]: DEBUG nova.network.neutron [-] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.355873] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68a75484-dbdf-4491-b8da-582a1a918d14 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.366602] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9cfcb23-5540-4f1d-aa4e-1281d4bf936e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.390021] env[62914]: DEBUG nova.compute.manager [req-d14621bc-b48e-40eb-be5b-47e16a5d0f15 req-68394a30-92b7-4cfb-8b8d-e1e91d0000b0 service nova] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Detach interface failed, port_id=1481dffb-497c-4588-8bb1-12ce0445c7c0, reason: Instance 58c2cc4a-153d-437b-8822-d3e389bd1a7f could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1083.490696] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: f80ee46d-9209-4b60-9e5a-f7c8764f4ccb] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.832057] env[62914]: INFO nova.compute.manager [-] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Took 1.62 seconds to deallocate network for instance. [ 1083.993998] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 29680cfb-8df3-4ff7-b6ee-0982577cc9cf] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.338129] env[62914]: DEBUG oslo_concurrency.lockutils [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.338418] env[62914]: DEBUG oslo_concurrency.lockutils [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.338656] env[62914]: DEBUG nova.objects.instance [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'resources' on Instance uuid 58c2cc4a-153d-437b-8822-d3e389bd1a7f {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1084.497150] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 43ee8948-8805-4d48-bd45-e93d2e2eb05d] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.002055] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 4bf76ce0-b5d2-4184-888c-d0ef39878356] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.022347] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-173e5014-35fa-4e07-abb1-e91701a075f2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.031149] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cbb1ef2-dc2c-4ad8-9434-4c9efd227266 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.061064] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f33004-887f-434b-bc54-1e8fdeed015f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.067738] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a860416-d496-4736-b950-adcb4cb95d70 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.080568] env[62914]: DEBUG nova.compute.provider_tree [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1085.505420] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: c7cf71d7-b823-471f-949f-9ef2b539d972] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.583131] env[62914]: DEBUG nova.scheduler.client.report [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1086.009093] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 1fe15423-f8be-4763-b55e-2cbb383ff01d] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.088084] env[62914]: DEBUG oslo_concurrency.lockutils [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.749s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.107118] env[62914]: INFO nova.scheduler.client.report [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Deleted allocations for instance 58c2cc4a-153d-437b-8822-d3e389bd1a7f [ 1086.512225] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 2274ca27-8e2d-435e-8570-97eb4ae3cc58] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.615282] env[62914]: DEBUG oslo_concurrency.lockutils [None req-37124ddd-802b-4ba2-8948-2857f2f07793 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "58c2cc4a-153d-437b-8822-d3e389bd1a7f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.513s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.016177] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: bd3e82ce-62e7-4761-bcd7-d1f835234b97] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.519620] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 93d420a1-6d8f-4919-a42f-55aebab853ae] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.634198] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.634460] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.634677] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.634864] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.635047] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.637064] env[62914]: INFO nova.compute.manager [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Terminating instance [ 1087.638829] env[62914]: DEBUG nova.compute.manager [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1087.639030] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1087.639863] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0c8f94-32ad-4b6b-9508-0e41bc8a914b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.647984] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1087.648243] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bea20e84-a9f5-4ba0-90b3-fd0d0ef4bd60 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.654365] env[62914]: DEBUG oslo_vmware.api [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1087.654365] env[62914]: value = "task-1352976" [ 1087.654365] env[62914]: _type = "Task" [ 1087.654365] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.662899] env[62914]: DEBUG oslo_vmware.api [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352976, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.022832] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 9577394d-1c73-4ed1-ba86-e7c246e32719] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.164476] env[62914]: DEBUG oslo_vmware.api [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352976, 'name': PowerOffVM_Task, 'duration_secs': 0.17559} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.164710] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1088.164879] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1088.165151] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4eae9667-5d74-47ea-add9-a146a228bc29 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.227234] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1088.227465] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1088.227643] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Deleting the datastore file [datastore1] 1028b121-ea93-4d9a-b2fd-0c2483c5c618 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1088.227908] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b4980ad0-c1be-4269-81ce-d778bff9538e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.234132] env[62914]: DEBUG oslo_vmware.api [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for the task: (returnval){ [ 1088.234132] env[62914]: value = "task-1352978" [ 1088.234132] env[62914]: _type = "Task" [ 1088.234132] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.241683] env[62914]: DEBUG oslo_vmware.api [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352978, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.525754] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: cb89c84f-414b-4dc7-9db5-12cbc30b52b4] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.743404] env[62914]: DEBUG oslo_vmware.api [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Task: {'id': task-1352978, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142695} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.743666] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1088.743856] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1088.744047] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1088.744249] env[62914]: INFO nova.compute.manager [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1088.744523] env[62914]: DEBUG oslo.service.loopingcall [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1088.744731] env[62914]: DEBUG nova.compute.manager [-] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1088.744827] env[62914]: DEBUG nova.network.neutron [-] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1088.977122] env[62914]: DEBUG nova.compute.manager [req-f00b86df-8a8a-4e50-9b65-63cb220b10d3 req-4a576a4e-6ac3-4ae8-b9a6-d338dc2889c3 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Received event network-vif-deleted-c240aa60-b896-4756-9de5-3965cfc8b21a {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1088.977304] env[62914]: INFO nova.compute.manager [req-f00b86df-8a8a-4e50-9b65-63cb220b10d3 req-4a576a4e-6ac3-4ae8-b9a6-d338dc2889c3 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Neutron deleted interface c240aa60-b896-4756-9de5-3965cfc8b21a; detaching it from the instance and deleting it from the info cache [ 1088.977472] env[62914]: DEBUG nova.network.neutron [req-f00b86df-8a8a-4e50-9b65-63cb220b10d3 req-4a576a4e-6ac3-4ae8-b9a6-d338dc2889c3 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.029212] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 46d139a1-a4ae-435e-9d6f-cfc06d706128] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.454820] env[62914]: DEBUG nova.network.neutron [-] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.480191] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0abb614f-ac5b-4c04-9347-d13ab053f8cc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.490431] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6acdd1af-f07f-4936-af46-ad0af4c879b1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.512903] env[62914]: DEBUG nova.compute.manager [req-f00b86df-8a8a-4e50-9b65-63cb220b10d3 req-4a576a4e-6ac3-4ae8-b9a6-d338dc2889c3 service nova] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Detach interface failed, port_id=c240aa60-b896-4756-9de5-3965cfc8b21a, reason: Instance 1028b121-ea93-4d9a-b2fd-0c2483c5c618 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1089.531854] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 9fec3e97-4203-4a90-8a43-600d11d7e7a0] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.957107] env[62914]: INFO nova.compute.manager [-] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Took 1.21 seconds to deallocate network for instance. [ 1090.034763] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 0f23eeda-7e0d-4b5f-9230-4ca1d5c8fd07] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.463523] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.463810] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.464072] env[62914]: DEBUG nova.objects.instance [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lazy-loading 'resources' on Instance uuid 1028b121-ea93-4d9a-b2fd-0c2483c5c618 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1090.537376] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 9fab500c-344c-46cd-b060-c8f1aa0f1cba] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.008548] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c114aaf8-f11c-48b9-a3dd-0484146cc7ed {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.015970] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fada87-433a-415f-896b-6ff38e79fcca {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.044454] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 87ef6b06-f699-4de3-8b89-854717074406] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.046682] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e07fbc-6d86-446f-babb-4d99ed8cf659 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.053677] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b734fd1-8790-4919-99c7-6deef4639123 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.066130] env[62914]: DEBUG nova.compute.provider_tree [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.549883] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 44339b45-5183-4882-aaec-4070adc3c3dd] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.569396] env[62914]: DEBUG nova.scheduler.client.report [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1092.053724] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.054060] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Cleaning up deleted instances with incomplete migration {{(pid=62914) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1092.074420] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.610s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.091232] env[62914]: INFO nova.scheduler.client.report [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Deleted allocations for instance 1028b121-ea93-4d9a-b2fd-0c2483c5c618 [ 1092.556758] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.598403] env[62914]: DEBUG oslo_concurrency.lockutils [None req-9dee2bb0-f510-4748-a734-b5c1862a3ae1 tempest-AttachInterfacesTestJSON-662165450 tempest-AttachInterfacesTestJSON-662165450-project-member] Lock "1028b121-ea93-4d9a-b2fd-0c2483c5c618" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.964s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.058593] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1094.563718] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1094.563718] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Starting heal instance info cache {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1094.563718] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Rebuilding the list of instances to heal {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1095.096285] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.096529] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquired lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.096562] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Forcefully refreshing network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1095.096723] env[62914]: DEBUG nova.objects.instance [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lazy-loading 'info_cache' on Instance uuid 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.846998] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.847368] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.847586] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.847787] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.847960] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.851716] env[62914]: INFO nova.compute.manager [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Terminating instance [ 1096.855021] env[62914]: DEBUG nova.compute.manager [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1096.855246] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1096.856128] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c596a26c-4569-4526-8163-53c9558bf3dd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.864309] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1096.868020] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ce4da2b-adf6-4c06-8803-3c20cf62aa11 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.868020] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Updating instance_info_cache with network_info: [{"id": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "address": "fa:16:3e:e8:d6:2b", "network": {"id": "74764d0b-64d3-4552-9517-ca2e87251c0b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1459372326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bb65abdffe44acdac0cdb8701e3e27b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape02166ec-c4", "ovs_interfaceid": "e02166ec-c4c0-44a5-a9d6-5927b987648d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.873553] env[62914]: DEBUG oslo_vmware.api [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1096.873553] env[62914]: value = "task-1352979" [ 1096.873553] env[62914]: _type = "Task" [ 1096.873553] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.883524] env[62914]: DEBUG oslo_vmware.api [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352979, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.372140] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Releasing lock "refresh_cache-062cd4d7-3cf1-4f31-9631-e9d1e194ad53" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.372396] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Updated the network info_cache for instance {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1097.372619] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.372774] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.372923] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.373076] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.373292] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._sync_power_states {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.384389] env[62914]: DEBUG oslo_vmware.api [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352979, 'name': PowerOffVM_Task, 'duration_secs': 0.173586} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.384654] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1097.384830] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1097.385099] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dcd1300a-0efc-4491-b994-b3e51ffde460 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.456601] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1097.456601] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1097.456601] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Deleting the datastore file [datastore2] 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1097.456601] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-02d7083e-db18-4398-8845-2c0a2caa146f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.465303] env[62914]: DEBUG oslo_vmware.api [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1097.465303] env[62914]: value = "task-1352981" [ 1097.465303] env[62914]: _type = "Task" [ 1097.465303] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.474577] env[62914]: DEBUG oslo_vmware.api [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352981, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.877199] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Getting list of instances from cluster (obj){ [ 1097.877199] env[62914]: value = "domain-c8" [ 1097.877199] env[62914]: _type = "ClusterComputeResource" [ 1097.877199] env[62914]: } {{(pid=62914) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1097.878292] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedd17aa-56e0-481a-9cf8-bb5db1fdcbcc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.887235] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Got total of 0 instances {{(pid=62914) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1097.887468] env[62914]: WARNING nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] While synchronizing instance power states, found 1 instances in the database and 0 instances on the hypervisor. [ 1097.887660] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Triggering sync for uuid 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 {{(pid=62914) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1097.888054] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.888335] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.888512] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62914) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1097.888717] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager.update_available_resource {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.975082] env[62914]: DEBUG oslo_vmware.api [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352981, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162783} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.975868] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1097.975868] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1097.975993] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1097.976090] env[62914]: INFO nova.compute.manager [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1097.976354] env[62914]: DEBUG oslo.service.loopingcall [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1097.976561] env[62914]: DEBUG nova.compute.manager [-] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1097.976659] env[62914]: DEBUG nova.network.neutron [-] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1098.325342] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "f1c5c877-9c29-4331-84f8-23b3c6b53725" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.325598] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "f1c5c877-9c29-4331-84f8-23b3c6b53725" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.384945] env[62914]: DEBUG nova.compute.manager [req-10d56ff6-77cb-40d5-bb66-8e619ae69947 req-ec31ed08-85f0-4fd1-bc8a-52368df29f27 service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Received event network-vif-deleted-e02166ec-c4c0-44a5-a9d6-5927b987648d {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1098.385058] env[62914]: INFO nova.compute.manager [req-10d56ff6-77cb-40d5-bb66-8e619ae69947 req-ec31ed08-85f0-4fd1-bc8a-52368df29f27 service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Neutron deleted interface e02166ec-c4c0-44a5-a9d6-5927b987648d; detaching it from the instance and deleting it from the info cache [ 1098.385246] env[62914]: DEBUG nova.network.neutron [req-10d56ff6-77cb-40d5-bb66-8e619ae69947 req-ec31ed08-85f0-4fd1-bc8a-52368df29f27 service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.391625] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.391867] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.392079] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.392261] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62914) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1098.393485] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4f805bc-a9e8-499b-93ac-b6ffa0c58f5d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.405490] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9913d83d-97c4-4b1b-906b-dcad90064004 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.422848] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d1f381-dbd7-423f-a892-90dc2042a95e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.429536] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c383a1f4-d74e-4351-bda3-aaad590c1365 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.464840] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181027MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62914) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1098.465041] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.465468] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.827730] env[62914]: DEBUG nova.compute.manager [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1098.860303] env[62914]: DEBUG nova.network.neutron [-] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.888495] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b7df3a1d-708d-4d74-a91b-55d2bea44943 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.897750] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a5c63f-01f0-4a28-bd1e-8074ce589ecf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.920049] env[62914]: DEBUG nova.compute.manager [req-10d56ff6-77cb-40d5-bb66-8e619ae69947 req-ec31ed08-85f0-4fd1-bc8a-52368df29f27 service nova] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Detach interface failed, port_id=e02166ec-c4c0-44a5-a9d6-5927b987648d, reason: Instance 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1099.349544] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.363173] env[62914]: INFO nova.compute.manager [-] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Took 1.39 seconds to deallocate network for instance. [ 1099.487148] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1099.869078] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.990321] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance f1c5c877-9c29-4331-84f8-23b3c6b53725 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1099.990665] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1099.990709] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1100.026148] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01fe4e7c-ca73-48db-8e23-eb21834589f3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.033998] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e3488ec-5462-4c30-bf59-e383c47de478 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.064330] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c62f35bb-7ac2-4439-a556-55f884c61f8d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.071102] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-133290d5-a613-4dbd-b334-0a8be1775185 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.083696] env[62914]: DEBUG nova.compute.provider_tree [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.586458] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1101.091761] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1101.092136] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.627s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.092666] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.743s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.093819] env[62914]: INFO nova.compute.claims [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1102.140581] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a9fe1ed-b030-49a0-a16e-81ec4eafa34e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.148240] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e2a4911-91b3-4ba6-aa24-3caf6692095d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.179155] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05570b62-8460-4b55-8338-cdf724da881f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.186804] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c673d8dd-b1fc-450b-9e6a-cd7eb9f1f763 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.200274] env[62914]: DEBUG nova.compute.provider_tree [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1102.703911] env[62914]: DEBUG nova.scheduler.client.report [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1103.209686] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.117s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.210326] env[62914]: DEBUG nova.compute.manager [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1103.213423] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.344s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.213643] env[62914]: DEBUG nova.objects.instance [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lazy-loading 'resources' on Instance uuid 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.716650] env[62914]: DEBUG nova.compute.utils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1103.718075] env[62914]: DEBUG nova.compute.manager [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1103.718267] env[62914]: DEBUG nova.network.neutron [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1103.756756] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6915d4d-47d1-40aa-8c5a-8758c3696a97 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.764116] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe581a76-3281-40b3-95a3-9d59823edeb3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.768367] env[62914]: DEBUG nova.policy [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7ab20b70ceb4f5bb700de50494399cc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbb45359291e4d699c21832d0265c1dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 1103.795362] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4089395c-b31e-4a99-971b-dd08149aca8b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.803881] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6a5b84-77b7-486e-b896-d47f20c649c1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.817749] env[62914]: DEBUG nova.compute.provider_tree [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.035576] env[62914]: DEBUG nova.network.neutron [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Successfully created port: e4f0d675-9d9c-4f76-bc5b-10ce73f445bf {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1104.224025] env[62914]: DEBUG nova.compute.manager [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1104.320986] env[62914]: DEBUG nova.scheduler.client.report [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1104.825507] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.612s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.846619] env[62914]: INFO nova.scheduler.client.report [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Deleted allocations for instance 062cd4d7-3cf1-4f31-9631-e9d1e194ad53 [ 1105.233658] env[62914]: DEBUG nova.compute.manager [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1105.259760] env[62914]: DEBUG nova.virt.hardware [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1105.260016] env[62914]: DEBUG nova.virt.hardware [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1105.260197] env[62914]: DEBUG nova.virt.hardware [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1105.260393] env[62914]: DEBUG nova.virt.hardware [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1105.260543] env[62914]: DEBUG nova.virt.hardware [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1105.260693] env[62914]: DEBUG nova.virt.hardware [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1105.260907] env[62914]: DEBUG nova.virt.hardware [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1105.261155] env[62914]: DEBUG nova.virt.hardware [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1105.261355] env[62914]: DEBUG nova.virt.hardware [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1105.261529] env[62914]: DEBUG nova.virt.hardware [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1105.261704] env[62914]: DEBUG nova.virt.hardware [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1105.262656] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b678b41-9fa8-4121-a286-b40df83a290a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.271843] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d76bd6a-ed3a-4072-9340-2281e8850908 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.355063] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bb350882-7ff0-4970-a4e3-f1962b7f7301 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.508s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.356037] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 7.468s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.356253] env[62914]: INFO nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] During sync_power_state the instance has a pending task (deleting). Skip. [ 1105.356437] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "062cd4d7-3cf1-4f31-9631-e9d1e194ad53" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.686511] env[62914]: DEBUG nova.compute.manager [req-a407e7bb-0b1e-47f2-9eb9-fdd41bb7edad req-0eb7cfe2-635b-448f-bdf2-3e98f80ed735 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Received event network-vif-plugged-e4f0d675-9d9c-4f76-bc5b-10ce73f445bf {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1105.686747] env[62914]: DEBUG oslo_concurrency.lockutils [req-a407e7bb-0b1e-47f2-9eb9-fdd41bb7edad req-0eb7cfe2-635b-448f-bdf2-3e98f80ed735 service nova] Acquiring lock "f1c5c877-9c29-4331-84f8-23b3c6b53725-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.686906] env[62914]: DEBUG oslo_concurrency.lockutils [req-a407e7bb-0b1e-47f2-9eb9-fdd41bb7edad req-0eb7cfe2-635b-448f-bdf2-3e98f80ed735 service nova] Lock "f1c5c877-9c29-4331-84f8-23b3c6b53725-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.687099] env[62914]: DEBUG oslo_concurrency.lockutils [req-a407e7bb-0b1e-47f2-9eb9-fdd41bb7edad req-0eb7cfe2-635b-448f-bdf2-3e98f80ed735 service nova] Lock "f1c5c877-9c29-4331-84f8-23b3c6b53725-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.687269] env[62914]: DEBUG nova.compute.manager [req-a407e7bb-0b1e-47f2-9eb9-fdd41bb7edad req-0eb7cfe2-635b-448f-bdf2-3e98f80ed735 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] No waiting events found dispatching network-vif-plugged-e4f0d675-9d9c-4f76-bc5b-10ce73f445bf {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1105.687487] env[62914]: WARNING nova.compute.manager [req-a407e7bb-0b1e-47f2-9eb9-fdd41bb7edad req-0eb7cfe2-635b-448f-bdf2-3e98f80ed735 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Received unexpected event network-vif-plugged-e4f0d675-9d9c-4f76-bc5b-10ce73f445bf for instance with vm_state building and task_state spawning. [ 1105.778791] env[62914]: DEBUG nova.network.neutron [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Successfully updated port: e4f0d675-9d9c-4f76-bc5b-10ce73f445bf {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1106.285406] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.285406] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.285406] env[62914]: DEBUG nova.network.neutron [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1106.824037] env[62914]: DEBUG nova.network.neutron [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1106.958906] env[62914]: DEBUG nova.network.neutron [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updating instance_info_cache with network_info: [{"id": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "address": "fa:16:3e:fb:87:48", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4f0d675-9d", "ovs_interfaceid": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.148963] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.149267] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.461500] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.461858] env[62914]: DEBUG nova.compute.manager [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Instance network_info: |[{"id": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "address": "fa:16:3e:fb:87:48", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4f0d675-9d", "ovs_interfaceid": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1107.462265] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:87:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '504a3211-8288-4b87-a52a-590466d80a49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e4f0d675-9d9c-4f76-bc5b-10ce73f445bf', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1107.469624] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Creating folder: Project (dbb45359291e4d699c21832d0265c1dd). Parent ref: group-v288131. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1107.469902] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c83c44dd-78eb-481e-9c04-e8697eb7e964 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.480541] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Created folder: Project (dbb45359291e4d699c21832d0265c1dd) in parent group-v288131. [ 1107.480726] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Creating folder: Instances. Parent ref: group-v288269. {{(pid=62914) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1107.480945] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c522c70f-b1c0-4527-bfc1-2f4f9ef4ac7b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.489474] env[62914]: INFO nova.virt.vmwareapi.vm_util [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Created folder: Instances in parent group-v288269. [ 1107.489692] env[62914]: DEBUG oslo.service.loopingcall [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1107.489870] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1107.490072] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da6a5cc9-c2bc-4fd6-a06b-c3cfc7c212d1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.507417] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1107.507417] env[62914]: value = "task-1352984" [ 1107.507417] env[62914]: _type = "Task" [ 1107.507417] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.514329] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352984, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.652411] env[62914]: DEBUG nova.compute.manager [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1107.713460] env[62914]: DEBUG nova.compute.manager [req-cabb21f8-ec02-4ed8-9379-2223d15028cd req-36a76fcc-8764-4ded-977c-cf74c1f84e47 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Received event network-changed-e4f0d675-9d9c-4f76-bc5b-10ce73f445bf {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1107.713692] env[62914]: DEBUG nova.compute.manager [req-cabb21f8-ec02-4ed8-9379-2223d15028cd req-36a76fcc-8764-4ded-977c-cf74c1f84e47 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Refreshing instance network info cache due to event network-changed-e4f0d675-9d9c-4f76-bc5b-10ce73f445bf. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1107.713939] env[62914]: DEBUG oslo_concurrency.lockutils [req-cabb21f8-ec02-4ed8-9379-2223d15028cd req-36a76fcc-8764-4ded-977c-cf74c1f84e47 service nova] Acquiring lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.714129] env[62914]: DEBUG oslo_concurrency.lockutils [req-cabb21f8-ec02-4ed8-9379-2223d15028cd req-36a76fcc-8764-4ded-977c-cf74c1f84e47 service nova] Acquired lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.714334] env[62914]: DEBUG nova.network.neutron [req-cabb21f8-ec02-4ed8-9379-2223d15028cd req-36a76fcc-8764-4ded-977c-cf74c1f84e47 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Refreshing network info cache for port e4f0d675-9d9c-4f76-bc5b-10ce73f445bf {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1108.017365] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352984, 'name': CreateVM_Task, 'duration_secs': 0.339911} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.017538] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1108.018206] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.018383] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.018709] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1108.018957] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a92791b-382c-4446-b591-af5213ee5660 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.023651] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1108.023651] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52bf9435-6e6a-5c20-ccce-f9abada23a44" [ 1108.023651] env[62914]: _type = "Task" [ 1108.023651] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.030729] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52bf9435-6e6a-5c20-ccce-f9abada23a44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.175788] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.176065] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.177722] env[62914]: INFO nova.compute.claims [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1108.411378] env[62914]: DEBUG nova.network.neutron [req-cabb21f8-ec02-4ed8-9379-2223d15028cd req-36a76fcc-8764-4ded-977c-cf74c1f84e47 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updated VIF entry in instance network info cache for port e4f0d675-9d9c-4f76-bc5b-10ce73f445bf. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1108.411766] env[62914]: DEBUG nova.network.neutron [req-cabb21f8-ec02-4ed8-9379-2223d15028cd req-36a76fcc-8764-4ded-977c-cf74c1f84e47 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updating instance_info_cache with network_info: [{"id": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "address": "fa:16:3e:fb:87:48", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4f0d675-9d", "ovs_interfaceid": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.534736] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52bf9435-6e6a-5c20-ccce-f9abada23a44, 'name': SearchDatastore_Task, 'duration_secs': 0.01074} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.535131] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.535295] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1108.535556] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.535728] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.535919] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1108.536234] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6566f38b-91c8-4781-902d-e28d1977ea07 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.548722] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1108.548955] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1108.549683] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a2c5529-2748-4c9c-bf17-b0aabbf0cac1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.555501] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1108.555501] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a94ba5-442e-27c9-2489-e6d31396c714" [ 1108.555501] env[62914]: _type = "Task" [ 1108.555501] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.564881] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a94ba5-442e-27c9-2489-e6d31396c714, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.914557] env[62914]: DEBUG oslo_concurrency.lockutils [req-cabb21f8-ec02-4ed8-9379-2223d15028cd req-36a76fcc-8764-4ded-977c-cf74c1f84e47 service nova] Releasing lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.069994] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a94ba5-442e-27c9-2489-e6d31396c714, 'name': SearchDatastore_Task, 'duration_secs': 0.038715} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.071111] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3485dfe-6fb5-4beb-a5a5-dd2d32a6a24a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.077632] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1109.077632] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]524fa211-6249-d6bc-bce5-1c1eb5bba7da" [ 1109.077632] env[62914]: _type = "Task" [ 1109.077632] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.087999] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]524fa211-6249-d6bc-bce5-1c1eb5bba7da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.229477] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a3d293-4214-49d7-bedd-4da133b75bf7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.236898] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad42e24e-6107-4c71-8327-3f72acda7a5f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.266821] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf0e92df-d60e-41f2-bf44-2f8bc9409651 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.274757] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24dfa6a7-82d4-4571-a34d-27c71c8f2c88 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.290495] env[62914]: DEBUG nova.compute.provider_tree [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1109.587986] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]524fa211-6249-d6bc-bce5-1c1eb5bba7da, 'name': SearchDatastore_Task, 'duration_secs': 0.040492} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.588449] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.588530] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] f1c5c877-9c29-4331-84f8-23b3c6b53725/f1c5c877-9c29-4331-84f8-23b3c6b53725.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1109.588792] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d6dfa465-2ffb-4e9d-a2a9-db420cefb6e3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.594917] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1109.594917] env[62914]: value = "task-1352985" [ 1109.594917] env[62914]: _type = "Task" [ 1109.594917] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.603717] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1352985, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.794581] env[62914]: DEBUG nova.scheduler.client.report [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1110.105095] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1352985, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.299468] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.123s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.300045] env[62914]: DEBUG nova.compute.manager [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1110.605782] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1352985, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.520447} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.606188] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] f1c5c877-9c29-4331-84f8-23b3c6b53725/f1c5c877-9c29-4331-84f8-23b3c6b53725.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1110.606285] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1110.606560] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d50353f-7374-40e0-9e8b-1a947cb8a7df {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.613214] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1110.613214] env[62914]: value = "task-1352986" [ 1110.613214] env[62914]: _type = "Task" [ 1110.613214] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.619877] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1352986, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.805867] env[62914]: DEBUG nova.compute.utils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1110.807342] env[62914]: DEBUG nova.compute.manager [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1110.807510] env[62914]: DEBUG nova.network.neutron [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1110.846676] env[62914]: DEBUG nova.policy [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '45afc180d8e441c78b16c28f6c18291b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bb65abdffe44acdac0cdb8701e3e27b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 1111.124261] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1352986, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096718} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.124261] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1111.124440] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372ccb3a-b606-4e6f-a50a-c67ace529adb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.152462] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] f1c5c877-9c29-4331-84f8-23b3c6b53725/f1c5c877-9c29-4331-84f8-23b3c6b53725.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1111.152462] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d460d22-fa59-4fba-a69c-e9e054ec9579 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.164161] env[62914]: DEBUG nova.network.neutron [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Successfully created port: 10301fa4-7a51-4bd7-8496-6e01598e68dc {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1111.168522] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1111.168522] env[62914]: value = "task-1352987" [ 1111.168522] env[62914]: _type = "Task" [ 1111.168522] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.176316] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1352987, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.311472] env[62914]: DEBUG nova.compute.manager [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1111.679258] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1352987, 'name': ReconfigVM_Task, 'duration_secs': 0.289255} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.679596] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Reconfigured VM instance instance-00000067 to attach disk [datastore2] f1c5c877-9c29-4331-84f8-23b3c6b53725/f1c5c877-9c29-4331-84f8-23b3c6b53725.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1111.680248] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d86adb8b-da3b-4a0e-bcf2-9b9d08f77c18 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.686056] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1111.686056] env[62914]: value = "task-1352988" [ 1111.686056] env[62914]: _type = "Task" [ 1111.686056] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.693336] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1352988, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.195965] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1352988, 'name': Rename_Task, 'duration_secs': 0.127516} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.196326] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1112.196579] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9e6e22ae-175e-4f03-a9a9-b129bf9e6450 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.201972] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1112.201972] env[62914]: value = "task-1352989" [ 1112.201972] env[62914]: _type = "Task" [ 1112.201972] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.211257] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1352989, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.321966] env[62914]: DEBUG nova.compute.manager [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1112.347177] env[62914]: DEBUG nova.virt.hardware [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1112.347446] env[62914]: DEBUG nova.virt.hardware [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1112.347610] env[62914]: DEBUG nova.virt.hardware [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1112.347794] env[62914]: DEBUG nova.virt.hardware [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1112.347943] env[62914]: DEBUG nova.virt.hardware [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1112.348113] env[62914]: DEBUG nova.virt.hardware [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1112.348338] env[62914]: DEBUG nova.virt.hardware [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1112.348489] env[62914]: DEBUG nova.virt.hardware [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1112.348655] env[62914]: DEBUG nova.virt.hardware [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1112.348817] env[62914]: DEBUG nova.virt.hardware [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1112.348989] env[62914]: DEBUG nova.virt.hardware [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1112.349856] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb7de2c-f10b-4928-9653-c52ac493ace9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.357471] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a987509a-c3c4-42fe-955e-a86d8d3b2bfd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.529027] env[62914]: DEBUG nova.compute.manager [req-f03abb49-0db5-4a7a-af67-31b36b1a1368 req-f6c64403-88dd-4be1-a633-4dc3abd71c89 service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Received event network-vif-plugged-10301fa4-7a51-4bd7-8496-6e01598e68dc {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.529310] env[62914]: DEBUG oslo_concurrency.lockutils [req-f03abb49-0db5-4a7a-af67-31b36b1a1368 req-f6c64403-88dd-4be1-a633-4dc3abd71c89 service nova] Acquiring lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.529560] env[62914]: DEBUG oslo_concurrency.lockutils [req-f03abb49-0db5-4a7a-af67-31b36b1a1368 req-f6c64403-88dd-4be1-a633-4dc3abd71c89 service nova] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.529781] env[62914]: DEBUG oslo_concurrency.lockutils [req-f03abb49-0db5-4a7a-af67-31b36b1a1368 req-f6c64403-88dd-4be1-a633-4dc3abd71c89 service nova] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.529998] env[62914]: DEBUG nova.compute.manager [req-f03abb49-0db5-4a7a-af67-31b36b1a1368 req-f6c64403-88dd-4be1-a633-4dc3abd71c89 service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] No waiting events found dispatching network-vif-plugged-10301fa4-7a51-4bd7-8496-6e01598e68dc {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1112.530231] env[62914]: WARNING nova.compute.manager [req-f03abb49-0db5-4a7a-af67-31b36b1a1368 req-f6c64403-88dd-4be1-a633-4dc3abd71c89 service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Received unexpected event network-vif-plugged-10301fa4-7a51-4bd7-8496-6e01598e68dc for instance with vm_state building and task_state spawning. [ 1112.611589] env[62914]: DEBUG nova.network.neutron [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Successfully updated port: 10301fa4-7a51-4bd7-8496-6e01598e68dc {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1112.711163] env[62914]: DEBUG oslo_vmware.api [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1352989, 'name': PowerOnVM_Task, 'duration_secs': 0.436037} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.711501] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1112.711610] env[62914]: INFO nova.compute.manager [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Took 7.48 seconds to spawn the instance on the hypervisor. [ 1112.711790] env[62914]: DEBUG nova.compute.manager [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1112.712537] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef365c5-d8a9-4734-9778-e20669726823 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.114316] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "refresh_cache-1005f4b0-e07a-4f32-9b3d-a2868a307bb9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.114491] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquired lock "refresh_cache-1005f4b0-e07a-4f32-9b3d-a2868a307bb9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.114660] env[62914]: DEBUG nova.network.neutron [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1113.228690] env[62914]: INFO nova.compute.manager [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Took 13.90 seconds to build instance. [ 1113.656120] env[62914]: DEBUG nova.network.neutron [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1113.731272] env[62914]: DEBUG oslo_concurrency.lockutils [None req-1dfcaee7-59a9-4e50-9f8e-55990a7bcc01 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "f1c5c877-9c29-4331-84f8-23b3c6b53725" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.405s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.876073] env[62914]: DEBUG nova.network.neutron [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Updating instance_info_cache with network_info: [{"id": "10301fa4-7a51-4bd7-8496-6e01598e68dc", "address": "fa:16:3e:49:56:0f", "network": {"id": "74764d0b-64d3-4552-9517-ca2e87251c0b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1459372326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bb65abdffe44acdac0cdb8701e3e27b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10301fa4-7a", "ovs_interfaceid": "10301fa4-7a51-4bd7-8496-6e01598e68dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.378502] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Releasing lock "refresh_cache-1005f4b0-e07a-4f32-9b3d-a2868a307bb9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.378877] env[62914]: DEBUG nova.compute.manager [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Instance network_info: |[{"id": "10301fa4-7a51-4bd7-8496-6e01598e68dc", "address": "fa:16:3e:49:56:0f", "network": {"id": "74764d0b-64d3-4552-9517-ca2e87251c0b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1459372326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bb65abdffe44acdac0cdb8701e3e27b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10301fa4-7a", "ovs_interfaceid": "10301fa4-7a51-4bd7-8496-6e01598e68dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1114.379378] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:56:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '31e77685-b4dd-4810-80ef-24115ea9ea62', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '10301fa4-7a51-4bd7-8496-6e01598e68dc', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1114.387304] env[62914]: DEBUG oslo.service.loopingcall [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1114.387524] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1114.387754] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b62092a8-af86-4df3-bd7b-909b3c2ebebb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.407785] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1114.407785] env[62914]: value = "task-1352990" [ 1114.407785] env[62914]: _type = "Task" [ 1114.407785] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.414903] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352990, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.557979] env[62914]: DEBUG nova.compute.manager [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Received event network-changed-10301fa4-7a51-4bd7-8496-6e01598e68dc {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1114.558255] env[62914]: DEBUG nova.compute.manager [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Refreshing instance network info cache due to event network-changed-10301fa4-7a51-4bd7-8496-6e01598e68dc. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1114.558458] env[62914]: DEBUG oslo_concurrency.lockutils [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] Acquiring lock "refresh_cache-1005f4b0-e07a-4f32-9b3d-a2868a307bb9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.558606] env[62914]: DEBUG oslo_concurrency.lockutils [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] Acquired lock "refresh_cache-1005f4b0-e07a-4f32-9b3d-a2868a307bb9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.558774] env[62914]: DEBUG nova.network.neutron [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Refreshing network info cache for port 10301fa4-7a51-4bd7-8496-6e01598e68dc {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1114.917566] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352990, 'name': CreateVM_Task, 'duration_secs': 0.395398} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.917866] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1114.918425] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.918598] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.918923] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1114.919183] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8226f98-07bc-4f31-99cf-9607a79a3076 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.923230] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1114.923230] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]520e8564-e43a-95aa-0828-8194fbb1ce76" [ 1114.923230] env[62914]: _type = "Task" [ 1114.923230] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.930158] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]520e8564-e43a-95aa-0828-8194fbb1ce76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.247959] env[62914]: DEBUG nova.network.neutron [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Updated VIF entry in instance network info cache for port 10301fa4-7a51-4bd7-8496-6e01598e68dc. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1115.248365] env[62914]: DEBUG nova.network.neutron [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Updating instance_info_cache with network_info: [{"id": "10301fa4-7a51-4bd7-8496-6e01598e68dc", "address": "fa:16:3e:49:56:0f", "network": {"id": "74764d0b-64d3-4552-9517-ca2e87251c0b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1459372326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bb65abdffe44acdac0cdb8701e3e27b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10301fa4-7a", "ovs_interfaceid": "10301fa4-7a51-4bd7-8496-6e01598e68dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.433378] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]520e8564-e43a-95aa-0828-8194fbb1ce76, 'name': SearchDatastore_Task, 'duration_secs': 0.009603} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.433693] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.433935] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1115.434202] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.434344] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.434528] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1115.434793] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-00d02482-f3e4-4e55-be20-61a1c146bb30 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.443016] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1115.443242] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1115.443916] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38eab054-fcbd-4210-8428-3efa295d3847 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.449156] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1115.449156] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a830aa-a01c-2623-6c86-8d7595bba287" [ 1115.449156] env[62914]: _type = "Task" [ 1115.449156] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.456268] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a830aa-a01c-2623-6c86-8d7595bba287, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.750872] env[62914]: DEBUG oslo_concurrency.lockutils [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] Releasing lock "refresh_cache-1005f4b0-e07a-4f32-9b3d-a2868a307bb9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.751168] env[62914]: DEBUG nova.compute.manager [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Received event network-changed-e4f0d675-9d9c-4f76-bc5b-10ce73f445bf {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1115.751440] env[62914]: DEBUG nova.compute.manager [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Refreshing instance network info cache due to event network-changed-e4f0d675-9d9c-4f76-bc5b-10ce73f445bf. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1115.751771] env[62914]: DEBUG oslo_concurrency.lockutils [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] Acquiring lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.752020] env[62914]: DEBUG oslo_concurrency.lockutils [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] Acquired lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.752256] env[62914]: DEBUG nova.network.neutron [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Refreshing network info cache for port e4f0d675-9d9c-4f76-bc5b-10ce73f445bf {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1115.959778] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52a830aa-a01c-2623-6c86-8d7595bba287, 'name': SearchDatastore_Task, 'duration_secs': 0.009837} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.960576] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3624a85-8646-4a2f-8046-4fe7592f9b68 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.965387] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1115.965387] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5258b855-daf7-0db5-a82d-826df360c7db" [ 1115.965387] env[62914]: _type = "Task" [ 1115.965387] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.972734] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5258b855-daf7-0db5-a82d-826df360c7db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.445293] env[62914]: DEBUG nova.network.neutron [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updated VIF entry in instance network info cache for port e4f0d675-9d9c-4f76-bc5b-10ce73f445bf. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1116.445651] env[62914]: DEBUG nova.network.neutron [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updating instance_info_cache with network_info: [{"id": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "address": "fa:16:3e:fb:87:48", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4f0d675-9d", "ovs_interfaceid": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.475065] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5258b855-daf7-0db5-a82d-826df360c7db, 'name': SearchDatastore_Task, 'duration_secs': 0.009154} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.475321] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.475570] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 1005f4b0-e07a-4f32-9b3d-a2868a307bb9/1005f4b0-e07a-4f32-9b3d-a2868a307bb9.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1116.475809] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-44c3e371-c563-497a-b80c-0cd90ff6849e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.481815] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1116.481815] env[62914]: value = "task-1352991" [ 1116.481815] env[62914]: _type = "Task" [ 1116.481815] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.489279] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352991, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.948132] env[62914]: DEBUG oslo_concurrency.lockutils [req-58455dbc-0bb2-40c6-8a16-06730aeef994 req-a12fe3dc-ad57-4bf2-bc70-72690dd90ee7 service nova] Releasing lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.991831] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352991, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443536} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.992212] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 1005f4b0-e07a-4f32-9b3d-a2868a307bb9/1005f4b0-e07a-4f32-9b3d-a2868a307bb9.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1116.992318] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1116.992593] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e3beb7fc-4fef-4ab3-9ccb-28d8e6d93496 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.998614] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1116.998614] env[62914]: value = "task-1352992" [ 1116.998614] env[62914]: _type = "Task" [ 1116.998614] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.007149] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352992, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.508353] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352992, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06234} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.508607] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1117.509388] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6282e40-21ee-4027-80c9-32fdf943de21 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.530437] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 1005f4b0-e07a-4f32-9b3d-a2868a307bb9/1005f4b0-e07a-4f32-9b3d-a2868a307bb9.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1117.530672] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18714cfe-e29a-4b35-9c7d-1889cd0ae069 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.548694] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1117.548694] env[62914]: value = "task-1352993" [ 1117.548694] env[62914]: _type = "Task" [ 1117.548694] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.560252] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352993, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.058964] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352993, 'name': ReconfigVM_Task, 'duration_secs': 0.299917} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.059359] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 1005f4b0-e07a-4f32-9b3d-a2868a307bb9/1005f4b0-e07a-4f32-9b3d-a2868a307bb9.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1118.059873] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8d92f243-31d7-429d-890b-42a7237c0e75 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.065833] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1118.065833] env[62914]: value = "task-1352994" [ 1118.065833] env[62914]: _type = "Task" [ 1118.065833] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.073015] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352994, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.575496] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352994, 'name': Rename_Task, 'duration_secs': 0.132462} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.575767] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1118.576061] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c10eaac4-1b46-4390-9dc9-f52b0bdb3d6b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.582596] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1118.582596] env[62914]: value = "task-1352995" [ 1118.582596] env[62914]: _type = "Task" [ 1118.582596] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.589577] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352995, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.092669] env[62914]: DEBUG oslo_vmware.api [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1352995, 'name': PowerOnVM_Task, 'duration_secs': 0.46898} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.093073] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1119.093136] env[62914]: INFO nova.compute.manager [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Took 6.77 seconds to spawn the instance on the hypervisor. [ 1119.093316] env[62914]: DEBUG nova.compute.manager [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1119.094107] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01df20b1-4959-4e62-bc8b-d74fb17e51d3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.610629] env[62914]: INFO nova.compute.manager [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Took 11.45 seconds to build instance. [ 1120.112241] env[62914]: DEBUG oslo_concurrency.lockutils [None req-c73bd754-16a7-4645-b06e-3207eee741cd tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.963s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.376355] env[62914]: DEBUG nova.compute.manager [req-3ca9c582-0b93-4a34-908b-cde8d2ab1408 req-b953782e-878a-42ee-9854-2dfbe8caec17 service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Received event network-changed-10301fa4-7a51-4bd7-8496-6e01598e68dc {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1120.376473] env[62914]: DEBUG nova.compute.manager [req-3ca9c582-0b93-4a34-908b-cde8d2ab1408 req-b953782e-878a-42ee-9854-2dfbe8caec17 service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Refreshing instance network info cache due to event network-changed-10301fa4-7a51-4bd7-8496-6e01598e68dc. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1120.376803] env[62914]: DEBUG oslo_concurrency.lockutils [req-3ca9c582-0b93-4a34-908b-cde8d2ab1408 req-b953782e-878a-42ee-9854-2dfbe8caec17 service nova] Acquiring lock "refresh_cache-1005f4b0-e07a-4f32-9b3d-a2868a307bb9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.376870] env[62914]: DEBUG oslo_concurrency.lockutils [req-3ca9c582-0b93-4a34-908b-cde8d2ab1408 req-b953782e-878a-42ee-9854-2dfbe8caec17 service nova] Acquired lock "refresh_cache-1005f4b0-e07a-4f32-9b3d-a2868a307bb9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.377214] env[62914]: DEBUG nova.network.neutron [req-3ca9c582-0b93-4a34-908b-cde8d2ab1408 req-b953782e-878a-42ee-9854-2dfbe8caec17 service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Refreshing network info cache for port 10301fa4-7a51-4bd7-8496-6e01598e68dc {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1121.129582] env[62914]: DEBUG nova.network.neutron [req-3ca9c582-0b93-4a34-908b-cde8d2ab1408 req-b953782e-878a-42ee-9854-2dfbe8caec17 service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Updated VIF entry in instance network info cache for port 10301fa4-7a51-4bd7-8496-6e01598e68dc. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1121.129970] env[62914]: DEBUG nova.network.neutron [req-3ca9c582-0b93-4a34-908b-cde8d2ab1408 req-b953782e-878a-42ee-9854-2dfbe8caec17 service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Updating instance_info_cache with network_info: [{"id": "10301fa4-7a51-4bd7-8496-6e01598e68dc", "address": "fa:16:3e:49:56:0f", "network": {"id": "74764d0b-64d3-4552-9517-ca2e87251c0b", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1459372326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bb65abdffe44acdac0cdb8701e3e27b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10301fa4-7a", "ovs_interfaceid": "10301fa4-7a51-4bd7-8496-6e01598e68dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.632905] env[62914]: DEBUG oslo_concurrency.lockutils [req-3ca9c582-0b93-4a34-908b-cde8d2ab1408 req-b953782e-878a-42ee-9854-2dfbe8caec17 service nova] Releasing lock "refresh_cache-1005f4b0-e07a-4f32-9b3d-a2868a307bb9" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.626492] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.626871] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.626871] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Starting heal instance info cache {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1136.626951] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Rebuilding the list of instances to heal {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1137.156161] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.157031] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquired lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.157031] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Forcefully refreshing network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1137.157031] env[62914]: DEBUG nova.objects.instance [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lazy-loading 'info_cache' on Instance uuid f1c5c877-9c29-4331-84f8-23b3c6b53725 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1138.875296] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updating instance_info_cache with network_info: [{"id": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "address": "fa:16:3e:fb:87:48", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4f0d675-9d", "ovs_interfaceid": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.377618] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Releasing lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.377769] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updated the network info_cache for instance {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1139.377968] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.378153] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.378300] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.378447] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.378586] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.378727] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.378853] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62914) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1139.378991] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager.update_available_resource {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.883055] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.883055] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.883055] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.883055] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62914) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1139.884334] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e5b27e-a006-491d-8925-adda9ac4e17d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.892385] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607a40e5-64e8-426b-a5d3-100a952daa0b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.905981] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0359eed2-4279-47f4-814e-4925ee01ceef {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.912010] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98464793-c65e-43ff-b5b9-c9b7cfe98bea {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.940732] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181027MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62914) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1139.940872] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.941074] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.964790] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance f1c5c877-9c29-4331-84f8-23b3c6b53725 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1140.965035] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 1005f4b0-e07a-4f32-9b3d-a2868a307bb9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1140.965131] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1140.965273] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1140.981519] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Refreshing inventories for resource provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1140.995118] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Updating ProviderTree inventory for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1140.995315] env[62914]: DEBUG nova.compute.provider_tree [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Updating inventory in ProviderTree for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1141.004877] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Refreshing aggregate associations for resource provider 8e1b3185-5ae3-4780-aee8-c507b157ad13, aggregates: None {{(pid=62914) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1141.022257] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Refreshing trait associations for resource provider 8e1b3185-5ae3-4780-aee8-c507b157ad13, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62914) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1141.051948] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef6632a-ab82-4ae7-b94a-8e6331d10ffa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.059148] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deab9cb3-e1f4-40cb-9823-1df2df463a74 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.088041] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51cbbcdf-1974-4781-9a17-ea7136cd8d74 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.095115] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f12f3f24-2c32-4aac-a471-dbcee7a4800f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.107645] env[62914]: DEBUG nova.compute.provider_tree [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.610876] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1142.117185] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1142.117465] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.176s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.905673] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "497c4c93-801e-41a6-ba8d-1c2f51710546" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.906288] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "497c4c93-801e-41a6-ba8d-1c2f51710546" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.409058] env[62914]: DEBUG nova.compute.manager [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1153.936238] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.936648] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.937944] env[62914]: INFO nova.compute.claims [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1154.991365] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40e8157-fbfb-4244-bb30-0f1a2d699d65 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.999078] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a32f35-7902-4051-bed6-bbd73d41a8e0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.028868] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdc313a-ac6a-45c3-8c32-6560518fdc55 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.035597] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe58fa3-2c53-486f-a2d4-a60a34f708fa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.048188] env[62914]: DEBUG nova.compute.provider_tree [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1155.551849] env[62914]: DEBUG nova.scheduler.client.report [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1156.057097] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.057713] env[62914]: DEBUG nova.compute.manager [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1156.563553] env[62914]: DEBUG nova.compute.utils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1156.564676] env[62914]: DEBUG nova.compute.manager [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1156.564848] env[62914]: DEBUG nova.network.neutron [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1156.617245] env[62914]: DEBUG nova.policy [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7ab20b70ceb4f5bb700de50494399cc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbb45359291e4d699c21832d0265c1dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 1156.867461] env[62914]: DEBUG nova.network.neutron [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Successfully created port: a05ccc42-a407-44c3-9987-b6493b36ad84 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1157.067988] env[62914]: DEBUG nova.compute.manager [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1157.628123] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.628389] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.078414] env[62914]: DEBUG nova.compute.manager [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1158.104595] env[62914]: DEBUG nova.virt.hardware [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1158.104921] env[62914]: DEBUG nova.virt.hardware [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1158.105103] env[62914]: DEBUG nova.virt.hardware [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1158.105296] env[62914]: DEBUG nova.virt.hardware [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1158.105449] env[62914]: DEBUG nova.virt.hardware [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1158.105618] env[62914]: DEBUG nova.virt.hardware [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1158.105871] env[62914]: DEBUG nova.virt.hardware [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1158.106055] env[62914]: DEBUG nova.virt.hardware [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1158.106232] env[62914]: DEBUG nova.virt.hardware [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1158.106397] env[62914]: DEBUG nova.virt.hardware [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1158.106583] env[62914]: DEBUG nova.virt.hardware [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1158.107443] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255886e6-7746-4d68-ace0-ade2a7286073 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.115755] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af4f756c-0cba-4fab-bd43-fbb9f199db2e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.131551] env[62914]: DEBUG nova.compute.utils [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1158.251337] env[62914]: DEBUG nova.compute.manager [req-5eb8ccaf-60a6-48c6-90a7-59b930e21d8f req-fbceaa47-cc66-4a99-b28a-592a7ef4ddc8 service nova] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Received event network-vif-plugged-a05ccc42-a407-44c3-9987-b6493b36ad84 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1158.251589] env[62914]: DEBUG oslo_concurrency.lockutils [req-5eb8ccaf-60a6-48c6-90a7-59b930e21d8f req-fbceaa47-cc66-4a99-b28a-592a7ef4ddc8 service nova] Acquiring lock "497c4c93-801e-41a6-ba8d-1c2f51710546-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.252188] env[62914]: DEBUG oslo_concurrency.lockutils [req-5eb8ccaf-60a6-48c6-90a7-59b930e21d8f req-fbceaa47-cc66-4a99-b28a-592a7ef4ddc8 service nova] Lock "497c4c93-801e-41a6-ba8d-1c2f51710546-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.252399] env[62914]: DEBUG oslo_concurrency.lockutils [req-5eb8ccaf-60a6-48c6-90a7-59b930e21d8f req-fbceaa47-cc66-4a99-b28a-592a7ef4ddc8 service nova] Lock "497c4c93-801e-41a6-ba8d-1c2f51710546-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.252549] env[62914]: DEBUG nova.compute.manager [req-5eb8ccaf-60a6-48c6-90a7-59b930e21d8f req-fbceaa47-cc66-4a99-b28a-592a7ef4ddc8 service nova] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] No waiting events found dispatching network-vif-plugged-a05ccc42-a407-44c3-9987-b6493b36ad84 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1158.252726] env[62914]: WARNING nova.compute.manager [req-5eb8ccaf-60a6-48c6-90a7-59b930e21d8f req-fbceaa47-cc66-4a99-b28a-592a7ef4ddc8 service nova] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Received unexpected event network-vif-plugged-a05ccc42-a407-44c3-9987-b6493b36ad84 for instance with vm_state building and task_state spawning. [ 1158.313845] env[62914]: DEBUG nova.network.neutron [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Successfully updated port: a05ccc42-a407-44c3-9987-b6493b36ad84 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1158.635551] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.818714] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "refresh_cache-497c4c93-801e-41a6-ba8d-1c2f51710546" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.818899] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "refresh_cache-497c4c93-801e-41a6-ba8d-1c2f51710546" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.818988] env[62914]: DEBUG nova.network.neutron [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1159.352729] env[62914]: DEBUG nova.network.neutron [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1159.471743] env[62914]: DEBUG nova.network.neutron [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Updating instance_info_cache with network_info: [{"id": "a05ccc42-a407-44c3-9987-b6493b36ad84", "address": "fa:16:3e:bb:9f:b1", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa05ccc42-a4", "ovs_interfaceid": "a05ccc42-a407-44c3-9987-b6493b36ad84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.698137] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.698421] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.698671] env[62914]: INFO nova.compute.manager [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Attaching volume b17c2ef2-b7c9-4904-beda-fd8ffdd44013 to /dev/sdb [ 1159.727986] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f01df1-7940-4597-b3f4-b7777c2290ad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.734755] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2f140b-fd9b-47eb-8501-35f523e848bd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.747111] env[62914]: DEBUG nova.virt.block_device [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Updating existing volume attachment record: 2bb574b8-2905-472c-9288-bef5e7939610 {{(pid=62914) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1159.974320] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "refresh_cache-497c4c93-801e-41a6-ba8d-1c2f51710546" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1159.974609] env[62914]: DEBUG nova.compute.manager [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Instance network_info: |[{"id": "a05ccc42-a407-44c3-9987-b6493b36ad84", "address": "fa:16:3e:bb:9f:b1", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa05ccc42-a4", "ovs_interfaceid": "a05ccc42-a407-44c3-9987-b6493b36ad84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1159.975128] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:9f:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '504a3211-8288-4b87-a52a-590466d80a49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a05ccc42-a407-44c3-9987-b6493b36ad84', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1159.982953] env[62914]: DEBUG oslo.service.loopingcall [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1159.983214] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1159.983507] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b5c839aa-07c8-4e1a-a6b8-bdc118d2fbdd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.005953] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1160.005953] env[62914]: value = "task-1352997" [ 1160.005953] env[62914]: _type = "Task" [ 1160.005953] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.013547] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352997, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.278691] env[62914]: DEBUG nova.compute.manager [req-468b20c5-e34f-4df4-b33f-162a09d06708 req-d04c8e1a-51f3-42dc-b67c-fab89fe2acc1 service nova] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Received event network-changed-a05ccc42-a407-44c3-9987-b6493b36ad84 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1160.278862] env[62914]: DEBUG nova.compute.manager [req-468b20c5-e34f-4df4-b33f-162a09d06708 req-d04c8e1a-51f3-42dc-b67c-fab89fe2acc1 service nova] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Refreshing instance network info cache due to event network-changed-a05ccc42-a407-44c3-9987-b6493b36ad84. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1160.279102] env[62914]: DEBUG oslo_concurrency.lockutils [req-468b20c5-e34f-4df4-b33f-162a09d06708 req-d04c8e1a-51f3-42dc-b67c-fab89fe2acc1 service nova] Acquiring lock "refresh_cache-497c4c93-801e-41a6-ba8d-1c2f51710546" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.279254] env[62914]: DEBUG oslo_concurrency.lockutils [req-468b20c5-e34f-4df4-b33f-162a09d06708 req-d04c8e1a-51f3-42dc-b67c-fab89fe2acc1 service nova] Acquired lock "refresh_cache-497c4c93-801e-41a6-ba8d-1c2f51710546" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.279439] env[62914]: DEBUG nova.network.neutron [req-468b20c5-e34f-4df4-b33f-162a09d06708 req-d04c8e1a-51f3-42dc-b67c-fab89fe2acc1 service nova] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Refreshing network info cache for port a05ccc42-a407-44c3-9987-b6493b36ad84 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1160.515773] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1352997, 'name': CreateVM_Task, 'duration_secs': 0.279229} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.516126] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1160.516601] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.516779] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.517137] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1160.517385] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed9623f3-234f-49b6-911a-95f78f6d7ebd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.521563] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1160.521563] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5204f230-a8f6-5fc9-3461-34c6882c7be1" [ 1160.521563] env[62914]: _type = "Task" [ 1160.521563] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.530058] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5204f230-a8f6-5fc9-3461-34c6882c7be1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.001121] env[62914]: DEBUG nova.network.neutron [req-468b20c5-e34f-4df4-b33f-162a09d06708 req-d04c8e1a-51f3-42dc-b67c-fab89fe2acc1 service nova] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Updated VIF entry in instance network info cache for port a05ccc42-a407-44c3-9987-b6493b36ad84. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1161.001489] env[62914]: DEBUG nova.network.neutron [req-468b20c5-e34f-4df4-b33f-162a09d06708 req-d04c8e1a-51f3-42dc-b67c-fab89fe2acc1 service nova] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Updating instance_info_cache with network_info: [{"id": "a05ccc42-a407-44c3-9987-b6493b36ad84", "address": "fa:16:3e:bb:9f:b1", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa05ccc42-a4", "ovs_interfaceid": "a05ccc42-a407-44c3-9987-b6493b36ad84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.032071] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5204f230-a8f6-5fc9-3461-34c6882c7be1, 'name': SearchDatastore_Task, 'duration_secs': 0.011311} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.032071] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.032071] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1161.032320] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1161.032320] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.032473] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1161.032725] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76ec02f5-002a-4098-ad33-0ccd8e4d7d80 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.040355] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1161.040538] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1161.041222] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd463d44-ffd9-4891-bd11-d6ee350eef8c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.046367] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1161.046367] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52050e2d-c281-a664-b775-6c14851e87ba" [ 1161.046367] env[62914]: _type = "Task" [ 1161.046367] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.053350] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52050e2d-c281-a664-b775-6c14851e87ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.504457] env[62914]: DEBUG oslo_concurrency.lockutils [req-468b20c5-e34f-4df4-b33f-162a09d06708 req-d04c8e1a-51f3-42dc-b67c-fab89fe2acc1 service nova] Releasing lock "refresh_cache-497c4c93-801e-41a6-ba8d-1c2f51710546" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.556720] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52050e2d-c281-a664-b775-6c14851e87ba, 'name': SearchDatastore_Task, 'duration_secs': 0.007539} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.557476] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e848f562-e607-472b-96b8-7afd51cd5677 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.562095] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1161.562095] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f45747-ecb6-4256-8192-dd35db160933" [ 1161.562095] env[62914]: _type = "Task" [ 1161.562095] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.568924] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f45747-ecb6-4256-8192-dd35db160933, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.072584] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52f45747-ecb6-4256-8192-dd35db160933, 'name': SearchDatastore_Task, 'duration_secs': 0.009035} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.072872] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1162.073181] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 497c4c93-801e-41a6-ba8d-1c2f51710546/497c4c93-801e-41a6-ba8d-1c2f51710546.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1162.073500] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e2b41245-5694-4fcd-8879-5de84c0b24ed {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.080956] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1162.080956] env[62914]: value = "task-1352999" [ 1162.080956] env[62914]: _type = "Task" [ 1162.080956] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.088507] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1352999, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.590581] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1352999, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.449395} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.590951] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore2] 497c4c93-801e-41a6-ba8d-1c2f51710546/497c4c93-801e-41a6-ba8d-1c2f51710546.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1162.591063] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1162.591411] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f6f05857-4070-4044-b023-c5d852e3cac2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.598353] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1162.598353] env[62914]: value = "task-1353000" [ 1162.598353] env[62914]: _type = "Task" [ 1162.598353] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.605187] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353000, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.108226] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353000, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070192} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.108470] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1163.109229] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2520374-4f06-4434-bc9d-779fc2b3f27a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.130135] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 497c4c93-801e-41a6-ba8d-1c2f51710546/497c4c93-801e-41a6-ba8d-1c2f51710546.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1163.130382] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af8ed097-a5f7-47ce-aea5-8524980970bb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.149653] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1163.149653] env[62914]: value = "task-1353001" [ 1163.149653] env[62914]: _type = "Task" [ 1163.149653] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.157030] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353001, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.661099] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353001, 'name': ReconfigVM_Task, 'duration_secs': 0.31443} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.661485] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 497c4c93-801e-41a6-ba8d-1c2f51710546/497c4c93-801e-41a6-ba8d-1c2f51710546.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1163.661971] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-758f1fd9-fefa-4759-b8f2-e5e06eb014de {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.667596] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1163.667596] env[62914]: value = "task-1353002" [ 1163.667596] env[62914]: _type = "Task" [ 1163.667596] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.674745] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353002, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.177304] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353002, 'name': Rename_Task, 'duration_secs': 0.133709} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.177578] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1164.177829] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eb0c82ab-7ecb-4d04-80dd-d8cd76a32a56 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.183516] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1164.183516] env[62914]: value = "task-1353003" [ 1164.183516] env[62914]: _type = "Task" [ 1164.183516] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.190510] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353003, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.290168] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Volume attach. Driver type: vmdk {{(pid=62914) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1164.290426] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288273', 'volume_id': 'b17c2ef2-b7c9-4904-beda-fd8ffdd44013', 'name': 'volume-b17c2ef2-b7c9-4904-beda-fd8ffdd44013', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1005f4b0-e07a-4f32-9b3d-a2868a307bb9', 'attached_at': '', 'detached_at': '', 'volume_id': 'b17c2ef2-b7c9-4904-beda-fd8ffdd44013', 'serial': 'b17c2ef2-b7c9-4904-beda-fd8ffdd44013'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1164.291327] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eade22a-a732-4326-9223-2d7bf7a6d853 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.307825] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-779d0c73-0ea6-4826-8598-37fcae8a741f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.331784] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] volume-b17c2ef2-b7c9-4904-beda-fd8ffdd44013/volume-b17c2ef2-b7c9-4904-beda-fd8ffdd44013.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1164.332088] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1a241bd-90c3-4855-9f47-ad5ce869235a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.350022] env[62914]: DEBUG oslo_vmware.api [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1164.350022] env[62914]: value = "task-1353004" [ 1164.350022] env[62914]: _type = "Task" [ 1164.350022] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.357640] env[62914]: DEBUG oslo_vmware.api [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353004, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.693384] env[62914]: DEBUG oslo_vmware.api [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353003, 'name': PowerOnVM_Task, 'duration_secs': 0.407839} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.693737] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1164.693867] env[62914]: INFO nova.compute.manager [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Took 6.62 seconds to spawn the instance on the hypervisor. [ 1164.694056] env[62914]: DEBUG nova.compute.manager [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1164.694852] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ca457f-eac7-4a65-9b02-760481227a3e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.859854] env[62914]: DEBUG oslo_vmware.api [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353004, 'name': ReconfigVM_Task, 'duration_secs': 0.33162} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.860143] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Reconfigured VM instance instance-00000068 to attach disk [datastore2] volume-b17c2ef2-b7c9-4904-beda-fd8ffdd44013/volume-b17c2ef2-b7c9-4904-beda-fd8ffdd44013.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1164.864805] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac3c0738-8fb8-4a85-ab8b-a106a8fdb492 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.879526] env[62914]: DEBUG oslo_vmware.api [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1164.879526] env[62914]: value = "task-1353005" [ 1164.879526] env[62914]: _type = "Task" [ 1164.879526] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.886722] env[62914]: DEBUG oslo_vmware.api [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353005, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.211099] env[62914]: INFO nova.compute.manager [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Took 11.30 seconds to build instance. [ 1165.389598] env[62914]: DEBUG oslo_vmware.api [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353005, 'name': ReconfigVM_Task, 'duration_secs': 0.198085} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.389919] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288273', 'volume_id': 'b17c2ef2-b7c9-4904-beda-fd8ffdd44013', 'name': 'volume-b17c2ef2-b7c9-4904-beda-fd8ffdd44013', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1005f4b0-e07a-4f32-9b3d-a2868a307bb9', 'attached_at': '', 'detached_at': '', 'volume_id': 'b17c2ef2-b7c9-4904-beda-fd8ffdd44013', 'serial': 'b17c2ef2-b7c9-4904-beda-fd8ffdd44013'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1165.713522] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7bcda9ee-44a3-41e5-99d0-4e30e10890ff tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "497c4c93-801e-41a6-ba8d-1c2f51710546" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.807s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.085139] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e087354f-ad24-4e4c-9383-b0339c4e7c8a tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "497c4c93-801e-41a6-ba8d-1c2f51710546" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.085416] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e087354f-ad24-4e4c-9383-b0339c4e7c8a tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "497c4c93-801e-41a6-ba8d-1c2f51710546" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.085612] env[62914]: DEBUG nova.compute.manager [None req-e087354f-ad24-4e4c-9383-b0339c4e7c8a tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1166.086557] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8f6af6-2330-4646-bc00-8b662f2375b8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.093587] env[62914]: DEBUG nova.compute.manager [None req-e087354f-ad24-4e4c-9383-b0339c4e7c8a tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62914) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1166.094138] env[62914]: DEBUG nova.objects.instance [None req-e087354f-ad24-4e4c-9383-b0339c4e7c8a tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lazy-loading 'flavor' on Instance uuid 497c4c93-801e-41a6-ba8d-1c2f51710546 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.426631] env[62914]: DEBUG nova.objects.instance [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lazy-loading 'flavor' on Instance uuid 1005f4b0-e07a-4f32-9b3d-a2868a307bb9 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.599207] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-e087354f-ad24-4e4c-9383-b0339c4e7c8a tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1166.599521] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a2db392-4903-43a3-8ee7-2ec767131086 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.607699] env[62914]: DEBUG oslo_vmware.api [None req-e087354f-ad24-4e4c-9383-b0339c4e7c8a tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1166.607699] env[62914]: value = "task-1353006" [ 1166.607699] env[62914]: _type = "Task" [ 1166.607699] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.615813] env[62914]: DEBUG oslo_vmware.api [None req-e087354f-ad24-4e4c-9383-b0339c4e7c8a tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353006, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.931740] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e0db32f3-ea8b-433f-b705-dc7eb63af58e tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.233s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.117630] env[62914]: DEBUG oslo_vmware.api [None req-e087354f-ad24-4e4c-9383-b0339c4e7c8a tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353006, 'name': PowerOffVM_Task, 'duration_secs': 0.188619} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.117893] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-e087354f-ad24-4e4c-9383-b0339c4e7c8a tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1167.118085] env[62914]: DEBUG nova.compute.manager [None req-e087354f-ad24-4e4c-9383-b0339c4e7c8a tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1167.118869] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6e9138-ddf0-4a0c-a577-837af8618f98 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.629890] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e087354f-ad24-4e4c-9383-b0339c4e7c8a tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "497c4c93-801e-41a6-ba8d-1c2f51710546" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.747678] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.747918] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.251486] env[62914]: DEBUG nova.compute.utils [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1168.506551] env[62914]: INFO nova.compute.manager [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Rebuilding instance [ 1168.558321] env[62914]: DEBUG nova.compute.manager [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1168.559662] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff07ca3-d9cf-43cf-b631-20994baf3d55 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.754685] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.074767] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1169.075036] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a495f2b7-bc75-4719-88c0-0f1e5ebeac75 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.082734] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1169.082734] env[62914]: value = "task-1353007" [ 1169.082734] env[62914]: _type = "Task" [ 1169.082734] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.090575] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353007, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.592637] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] VM already powered off {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1169.592994] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1169.593569] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d01fe2-9050-4e21-8df3-33bd6b392c76 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.599992] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1169.600228] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5eccdbad-99a9-49e7-98c3-a678767bef15 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.675015] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1169.675313] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1169.675564] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleting the datastore file [datastore2] 497c4c93-801e-41a6-ba8d-1c2f51710546 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1169.675907] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba7acedd-991d-4093-bba4-081bb49d44db {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.682195] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1169.682195] env[62914]: value = "task-1353009" [ 1169.682195] env[62914]: _type = "Task" [ 1169.682195] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.689867] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353009, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.810202] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.810551] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.810863] env[62914]: INFO nova.compute.manager [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Attaching volume f08700c6-2223-4dab-8faf-8da66f8e8250 to /dev/sdc [ 1169.846548] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b9818f8-2999-4c98-a8ef-761d7f05ad55 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.853816] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adaecb7c-f407-4cf2-b66f-559cf512df82 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.866956] env[62914]: DEBUG nova.virt.block_device [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Updating existing volume attachment record: 02825440-9082-4573-aada-8d02332df7fe {{(pid=62914) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1170.192579] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353009, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165859} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.192841] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1170.193039] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1170.193223] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1171.229064] env[62914]: DEBUG nova.virt.hardware [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1171.229335] env[62914]: DEBUG nova.virt.hardware [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1171.229489] env[62914]: DEBUG nova.virt.hardware [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1171.229677] env[62914]: DEBUG nova.virt.hardware [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1171.229827] env[62914]: DEBUG nova.virt.hardware [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1171.229977] env[62914]: DEBUG nova.virt.hardware [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1171.230224] env[62914]: DEBUG nova.virt.hardware [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1171.230365] env[62914]: DEBUG nova.virt.hardware [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1171.230534] env[62914]: DEBUG nova.virt.hardware [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1171.230694] env[62914]: DEBUG nova.virt.hardware [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1171.230866] env[62914]: DEBUG nova.virt.hardware [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1171.231764] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec65b0d-fc31-4367-bd9a-42218bd8771e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.239225] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3b1573-aa0c-4f46-ad7d-780166421be4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.252403] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:9f:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '504a3211-8288-4b87-a52a-590466d80a49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a05ccc42-a407-44c3-9987-b6493b36ad84', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1171.259567] env[62914]: DEBUG oslo.service.loopingcall [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.259803] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1171.259998] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f3fe8142-ce78-4c25-9a1d-7545bbd911cb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.278366] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1171.278366] env[62914]: value = "task-1353011" [ 1171.278366] env[62914]: _type = "Task" [ 1171.278366] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.285623] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1353011, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.788467] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1353011, 'name': CreateVM_Task, 'duration_secs': 0.287775} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.788711] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1171.789440] env[62914]: DEBUG oslo_concurrency.lockutils [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1171.789610] env[62914]: DEBUG oslo_concurrency.lockutils [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.789934] env[62914]: DEBUG oslo_concurrency.lockutils [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1171.790201] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b82ef90-5f71-42bd-ade8-ebf9d255436f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.794506] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1171.794506] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52471af5-77ee-ff83-f5fb-233b2210fd98" [ 1171.794506] env[62914]: _type = "Task" [ 1171.794506] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.801597] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52471af5-77ee-ff83-f5fb-233b2210fd98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.304633] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52471af5-77ee-ff83-f5fb-233b2210fd98, 'name': SearchDatastore_Task, 'duration_secs': 0.009457} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.305033] env[62914]: DEBUG oslo_concurrency.lockutils [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1172.305121] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1172.305359] env[62914]: DEBUG oslo_concurrency.lockutils [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1172.305509] env[62914]: DEBUG oslo_concurrency.lockutils [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.305698] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1172.305990] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dc0a4a5e-f231-4965-9d55-3c9c1cac9513 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.313618] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1172.313825] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1172.314493] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94434667-6965-449b-97eb-5d1eba6f5a50 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.319032] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1172.319032] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5240f1fa-934a-07c1-b692-986df405935d" [ 1172.319032] env[62914]: _type = "Task" [ 1172.319032] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.326222] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5240f1fa-934a-07c1-b692-986df405935d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.828841] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5240f1fa-934a-07c1-b692-986df405935d, 'name': SearchDatastore_Task, 'duration_secs': 0.008258} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.829618] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-285ea781-6e37-47ef-9be2-705b61e8cbb3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.834613] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1172.834613] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5284b4c7-e649-66e0-8020-307a485b33e3" [ 1172.834613] env[62914]: _type = "Task" [ 1172.834613] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.841690] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5284b4c7-e649-66e0-8020-307a485b33e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.345234] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5284b4c7-e649-66e0-8020-307a485b33e3, 'name': SearchDatastore_Task, 'duration_secs': 0.008696} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.345577] env[62914]: DEBUG oslo_concurrency.lockutils [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1173.345757] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 497c4c93-801e-41a6-ba8d-1c2f51710546/497c4c93-801e-41a6-ba8d-1c2f51710546.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1173.346036] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0d2d38ca-c6c9-4b67-9d2a-26805f0eafaa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.352329] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1173.352329] env[62914]: value = "task-1353013" [ 1173.352329] env[62914]: _type = "Task" [ 1173.352329] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.359184] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353013, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.861813] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353013, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.433062} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.862077] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 497c4c93-801e-41a6-ba8d-1c2f51710546/497c4c93-801e-41a6-ba8d-1c2f51710546.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1173.862322] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1173.862573] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5ca6711a-ee50-44d9-ae51-59db967535cf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.868136] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1173.868136] env[62914]: value = "task-1353014" [ 1173.868136] env[62914]: _type = "Task" [ 1173.868136] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.875092] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353014, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.378020] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353014, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069412} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.378362] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1174.378978] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865b7bf8-b5d3-4e33-9418-a2b88a865f63 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.400836] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 497c4c93-801e-41a6-ba8d-1c2f51710546/497c4c93-801e-41a6-ba8d-1c2f51710546.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1174.401087] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99c3c3a9-3577-423b-8bd9-d82759fd2108 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.415608] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Volume attach. Driver type: vmdk {{(pid=62914) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1174.415827] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288275', 'volume_id': 'f08700c6-2223-4dab-8faf-8da66f8e8250', 'name': 'volume-f08700c6-2223-4dab-8faf-8da66f8e8250', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1005f4b0-e07a-4f32-9b3d-a2868a307bb9', 'attached_at': '', 'detached_at': '', 'volume_id': 'f08700c6-2223-4dab-8faf-8da66f8e8250', 'serial': 'f08700c6-2223-4dab-8faf-8da66f8e8250'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1174.416581] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-571144c7-d1b5-4220-9a91-56220b5fef23 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.433325] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87a8f2e5-42de-4801-beed-332fdf2f8b79 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.435710] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1174.435710] env[62914]: value = "task-1353015" [ 1174.435710] env[62914]: _type = "Task" [ 1174.435710] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.461149] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] volume-f08700c6-2223-4dab-8faf-8da66f8e8250/volume-f08700c6-2223-4dab-8faf-8da66f8e8250.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1174.461787] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f72d1275-1e64-48f3-983a-3d39e0bc85d7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.477068] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353015, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.481697] env[62914]: DEBUG oslo_vmware.api [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1174.481697] env[62914]: value = "task-1353016" [ 1174.481697] env[62914]: _type = "Task" [ 1174.481697] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.488867] env[62914]: DEBUG oslo_vmware.api [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353016, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.944677] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353015, 'name': ReconfigVM_Task, 'duration_secs': 0.286649} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.944968] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 497c4c93-801e-41a6-ba8d-1c2f51710546/497c4c93-801e-41a6-ba8d-1c2f51710546.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1174.945713] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0604072-4763-4734-8372-7d3df6113705 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.952332] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1174.952332] env[62914]: value = "task-1353017" [ 1174.952332] env[62914]: _type = "Task" [ 1174.952332] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.960128] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353017, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.990678] env[62914]: DEBUG oslo_vmware.api [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353016, 'name': ReconfigVM_Task, 'duration_secs': 0.358037} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.990992] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Reconfigured VM instance instance-00000068 to attach disk [datastore1] volume-f08700c6-2223-4dab-8faf-8da66f8e8250/volume-f08700c6-2223-4dab-8faf-8da66f8e8250.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1174.996456] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7be569e-12bf-4a94-a686-16bb8650f9fe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.011345] env[62914]: DEBUG oslo_vmware.api [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1175.011345] env[62914]: value = "task-1353018" [ 1175.011345] env[62914]: _type = "Task" [ 1175.011345] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.018901] env[62914]: DEBUG oslo_vmware.api [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353018, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.462471] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353017, 'name': Rename_Task, 'duration_secs': 0.136458} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.462850] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1175.463029] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-431378cf-6406-44be-946b-a94ace482793 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.469078] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1175.469078] env[62914]: value = "task-1353019" [ 1175.469078] env[62914]: _type = "Task" [ 1175.469078] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.476174] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353019, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.520151] env[62914]: DEBUG oslo_vmware.api [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353018, 'name': ReconfigVM_Task, 'duration_secs': 0.137747} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.520459] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288275', 'volume_id': 'f08700c6-2223-4dab-8faf-8da66f8e8250', 'name': 'volume-f08700c6-2223-4dab-8faf-8da66f8e8250', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1005f4b0-e07a-4f32-9b3d-a2868a307bb9', 'attached_at': '', 'detached_at': '', 'volume_id': 'f08700c6-2223-4dab-8faf-8da66f8e8250', 'serial': 'f08700c6-2223-4dab-8faf-8da66f8e8250'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1175.980438] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353019, 'name': PowerOnVM_Task, 'duration_secs': 0.402462} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.980714] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1175.980921] env[62914]: DEBUG nova.compute.manager [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1175.981730] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a025cf9-d917-4f24-91ff-c30d02e70ff5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.492502] env[62914]: INFO nova.compute.manager [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] bringing vm to original state: 'stopped' [ 1176.555771] env[62914]: DEBUG nova.objects.instance [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lazy-loading 'flavor' on Instance uuid 1005f4b0-e07a-4f32-9b3d-a2868a307bb9 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1177.061224] env[62914]: DEBUG oslo_concurrency.lockutils [None req-7d17d4a1-f175-447e-b8af-258089537fb0 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.251s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.333238] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.333486] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.500193] env[62914]: DEBUG oslo_concurrency.lockutils [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "497c4c93-801e-41a6-ba8d-1c2f51710546" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.500524] env[62914]: DEBUG oslo_concurrency.lockutils [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "497c4c93-801e-41a6-ba8d-1c2f51710546" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.500671] env[62914]: DEBUG nova.compute.manager [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1177.501636] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0166bd44-328c-4397-b7bc-143528ebd2f0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.508430] env[62914]: DEBUG nova.compute.manager [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62914) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1177.510387] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1177.510617] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8bcef7da-b23f-419a-a127-cda85a1fee5d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.517142] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1177.517142] env[62914]: value = "task-1353020" [ 1177.517142] env[62914]: _type = "Task" [ 1177.517142] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.524278] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353020, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.836609] env[62914]: INFO nova.compute.manager [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Detaching volume b17c2ef2-b7c9-4904-beda-fd8ffdd44013 [ 1177.870395] env[62914]: INFO nova.virt.block_device [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Attempting to driver detach volume b17c2ef2-b7c9-4904-beda-fd8ffdd44013 from mountpoint /dev/sdb [ 1177.870642] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Volume detach. Driver type: vmdk {{(pid=62914) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1177.870834] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288273', 'volume_id': 'b17c2ef2-b7c9-4904-beda-fd8ffdd44013', 'name': 'volume-b17c2ef2-b7c9-4904-beda-fd8ffdd44013', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1005f4b0-e07a-4f32-9b3d-a2868a307bb9', 'attached_at': '', 'detached_at': '', 'volume_id': 'b17c2ef2-b7c9-4904-beda-fd8ffdd44013', 'serial': 'b17c2ef2-b7c9-4904-beda-fd8ffdd44013'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1177.871792] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46d1666a-1ecb-4baf-97b9-dc254ab91d29 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.896347] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-952896f8-ccd5-4d34-a285-01c9da64af0c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.902943] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bed780c-e70a-4933-8efb-3f49b159b340 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.925073] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99db25c5-6b1b-4b7b-b1af-dc3a0ee3e10a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.938900] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] The volume has not been displaced from its original location: [datastore2] volume-b17c2ef2-b7c9-4904-beda-fd8ffdd44013/volume-b17c2ef2-b7c9-4904-beda-fd8ffdd44013.vmdk. No consolidation needed. {{(pid=62914) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1177.944170] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Reconfiguring VM instance instance-00000068 to detach disk 2001 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1177.944434] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a72ee2db-194f-45f4-a164-31b55d478fc9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.962016] env[62914]: DEBUG oslo_vmware.api [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1177.962016] env[62914]: value = "task-1353021" [ 1177.962016] env[62914]: _type = "Task" [ 1177.962016] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.971454] env[62914]: DEBUG oslo_vmware.api [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353021, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.026670] env[62914]: DEBUG oslo_vmware.api [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353020, 'name': PowerOffVM_Task, 'duration_secs': 0.142347} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.026942] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1178.027136] env[62914]: DEBUG nova.compute.manager [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1178.027880] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15b4b81-bf29-4d44-95e8-0153a2e576d1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.471883] env[62914]: DEBUG oslo_vmware.api [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353021, 'name': ReconfigVM_Task, 'duration_secs': 0.212222} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.472175] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Reconfigured VM instance instance-00000068 to detach disk 2001 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1178.476786] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83987495-d3af-4e76-a0f8-1c55233984dc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.491308] env[62914]: DEBUG oslo_vmware.api [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1178.491308] env[62914]: value = "task-1353022" [ 1178.491308] env[62914]: _type = "Task" [ 1178.491308] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.498622] env[62914]: DEBUG oslo_vmware.api [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353022, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.540174] env[62914]: DEBUG oslo_concurrency.lockutils [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "497c4c93-801e-41a6-ba8d-1c2f51710546" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.040s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.001283] env[62914]: DEBUG oslo_vmware.api [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353022, 'name': ReconfigVM_Task, 'duration_secs': 0.130411} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.001662] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288273', 'volume_id': 'b17c2ef2-b7c9-4904-beda-fd8ffdd44013', 'name': 'volume-b17c2ef2-b7c9-4904-beda-fd8ffdd44013', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1005f4b0-e07a-4f32-9b3d-a2868a307bb9', 'attached_at': '', 'detached_at': '', 'volume_id': 'b17c2ef2-b7c9-4904-beda-fd8ffdd44013', 'serial': 'b17c2ef2-b7c9-4904-beda-fd8ffdd44013'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1179.047098] env[62914]: DEBUG oslo_concurrency.lockutils [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.047358] env[62914]: DEBUG oslo_concurrency.lockutils [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.047578] env[62914]: DEBUG nova.objects.instance [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62914) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1179.540896] env[62914]: DEBUG nova.objects.instance [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lazy-loading 'flavor' on Instance uuid 1005f4b0-e07a-4f32-9b3d-a2868a307bb9 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1180.055910] env[62914]: DEBUG oslo_concurrency.lockutils [None req-78a4120f-9655-42f8-b27a-d3274d3919ae tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.008s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.109206] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "497c4c93-801e-41a6-ba8d-1c2f51710546" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1180.109639] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "497c4c93-801e-41a6-ba8d-1c2f51710546" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1180.109962] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "497c4c93-801e-41a6-ba8d-1c2f51710546-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1180.110325] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "497c4c93-801e-41a6-ba8d-1c2f51710546-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1180.110602] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "497c4c93-801e-41a6-ba8d-1c2f51710546-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.113443] env[62914]: INFO nova.compute.manager [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Terminating instance [ 1180.115984] env[62914]: DEBUG nova.compute.manager [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1180.116325] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1180.117638] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c17c39-814b-4e7c-a5e3-2b84187e1773 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.125461] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1180.125700] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f437487b-e2ca-4e14-b2b7-c5e1a5ad12b2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.190591] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1180.190591] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1180.190591] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleting the datastore file [datastore1] 497c4c93-801e-41a6-ba8d-1c2f51710546 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1180.190910] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5fbd021-fd93-4acf-b67a-7a62bf5cafb2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.197903] env[62914]: DEBUG oslo_vmware.api [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1180.197903] env[62914]: value = "task-1353024" [ 1180.197903] env[62914]: _type = "Task" [ 1180.197903] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.205920] env[62914]: DEBUG oslo_vmware.api [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353024, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.549276] env[62914]: DEBUG oslo_concurrency.lockutils [None req-13426823-d493-418e-a966-b3a0f2284b4b tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.215s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.561200] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1180.561470] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1180.709416] env[62914]: DEBUG oslo_vmware.api [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353024, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133807} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.709689] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1180.709883] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1180.710078] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1180.710264] env[62914]: INFO nova.compute.manager [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1180.710515] env[62914]: DEBUG oslo.service.loopingcall [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1180.710715] env[62914]: DEBUG nova.compute.manager [-] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1180.710809] env[62914]: DEBUG nova.network.neutron [-] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1181.027457] env[62914]: DEBUG nova.compute.manager [req-18553143-aa7b-4d84-9b0e-77dbeb55d992 req-8812bc74-f894-49a7-888d-1b70a312ae0d service nova] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Received event network-vif-deleted-a05ccc42-a407-44c3-9987-b6493b36ad84 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1181.027677] env[62914]: INFO nova.compute.manager [req-18553143-aa7b-4d84-9b0e-77dbeb55d992 req-8812bc74-f894-49a7-888d-1b70a312ae0d service nova] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Neutron deleted interface a05ccc42-a407-44c3-9987-b6493b36ad84; detaching it from the instance and deleting it from the info cache [ 1181.027855] env[62914]: DEBUG nova.network.neutron [req-18553143-aa7b-4d84-9b0e-77dbeb55d992 req-8812bc74-f894-49a7-888d-1b70a312ae0d service nova] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.064636] env[62914]: INFO nova.compute.manager [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Detaching volume f08700c6-2223-4dab-8faf-8da66f8e8250 [ 1181.096510] env[62914]: INFO nova.virt.block_device [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Attempting to driver detach volume f08700c6-2223-4dab-8faf-8da66f8e8250 from mountpoint /dev/sdc [ 1181.096758] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Volume detach. Driver type: vmdk {{(pid=62914) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1181.096946] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288275', 'volume_id': 'f08700c6-2223-4dab-8faf-8da66f8e8250', 'name': 'volume-f08700c6-2223-4dab-8faf-8da66f8e8250', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1005f4b0-e07a-4f32-9b3d-a2868a307bb9', 'attached_at': '', 'detached_at': '', 'volume_id': 'f08700c6-2223-4dab-8faf-8da66f8e8250', 'serial': 'f08700c6-2223-4dab-8faf-8da66f8e8250'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1181.097832] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f26a6374-28b9-4547-be49-3b2fbc7c5e4a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.120568] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7afad841-3a2d-4089-bf3a-613ef0cc3a78 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.128034] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee0664b-aef2-4561-8989-783bdcd92544 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.147508] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c386a4f-5cb5-42b4-88b5-0402e4a4600f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.161650] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] The volume has not been displaced from its original location: [datastore1] volume-f08700c6-2223-4dab-8faf-8da66f8e8250/volume-f08700c6-2223-4dab-8faf-8da66f8e8250.vmdk. No consolidation needed. {{(pid=62914) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1181.166763] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Reconfiguring VM instance instance-00000068 to detach disk 2002 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1181.167031] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d09f7b5-b079-4857-94e4-3557336fd700 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.185190] env[62914]: DEBUG oslo_vmware.api [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1181.185190] env[62914]: value = "task-1353025" [ 1181.185190] env[62914]: _type = "Task" [ 1181.185190] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.192551] env[62914]: DEBUG oslo_vmware.api [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353025, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.502282] env[62914]: DEBUG nova.network.neutron [-] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.530152] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e5a623e-438b-41c5-99cc-4a09a8a02c4a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.539205] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56c00a8-f059-4f67-b0a9-98e0ac1c1603 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.562976] env[62914]: DEBUG nova.compute.manager [req-18553143-aa7b-4d84-9b0e-77dbeb55d992 req-8812bc74-f894-49a7-888d-1b70a312ae0d service nova] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Detach interface failed, port_id=a05ccc42-a407-44c3-9987-b6493b36ad84, reason: Instance 497c4c93-801e-41a6-ba8d-1c2f51710546 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1181.697067] env[62914]: DEBUG oslo_vmware.api [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353025, 'name': ReconfigVM_Task, 'duration_secs': 0.192793} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.697447] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Reconfigured VM instance instance-00000068 to detach disk 2002 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1181.702038] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b42d6798-4d99-4b25-ab5a-71443087a4dc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.716391] env[62914]: DEBUG oslo_vmware.api [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1181.716391] env[62914]: value = "task-1353026" [ 1181.716391] env[62914]: _type = "Task" [ 1181.716391] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.723635] env[62914]: DEBUG oslo_vmware.api [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353026, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.005111] env[62914]: INFO nova.compute.manager [-] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Took 1.29 seconds to deallocate network for instance. [ 1182.226147] env[62914]: DEBUG oslo_vmware.api [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353026, 'name': ReconfigVM_Task, 'duration_secs': 0.121078} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.226464] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288275', 'volume_id': 'f08700c6-2223-4dab-8faf-8da66f8e8250', 'name': 'volume-f08700c6-2223-4dab-8faf-8da66f8e8250', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1005f4b0-e07a-4f32-9b3d-a2868a307bb9', 'attached_at': '', 'detached_at': '', 'volume_id': 'f08700c6-2223-4dab-8faf-8da66f8e8250', 'serial': 'f08700c6-2223-4dab-8faf-8da66f8e8250'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1182.512452] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.512816] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.513020] env[62914]: DEBUG nova.objects.instance [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lazy-loading 'resources' on Instance uuid 497c4c93-801e-41a6-ba8d-1c2f51710546 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1182.765541] env[62914]: DEBUG nova.objects.instance [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lazy-loading 'flavor' on Instance uuid 1005f4b0-e07a-4f32-9b3d-a2868a307bb9 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1183.067041] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b08adc17-f49a-428f-a531-dc87a4258f57 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.074622] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b71b9c-29da-4635-bb75-3b14cfaa5bdf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.105288] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2677a8d4-6673-4584-8b2f-0aa64e15e236 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.112804] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-442a7f4d-1ad1-4499-986f-5346a8a02889 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.125771] env[62914]: DEBUG nova.compute.provider_tree [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1183.629517] env[62914]: DEBUG nova.scheduler.client.report [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1183.772664] env[62914]: DEBUG oslo_concurrency.lockutils [None req-a51246c2-93c6-4578-9b25-6c9eedd78698 tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.211s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.134542] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.622s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.157738] env[62914]: INFO nova.scheduler.client.report [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleted allocations for instance 497c4c93-801e-41a6-ba8d-1c2f51710546 [ 1184.664903] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e1b57359-e82a-4416-aedd-2ca85bdee17b tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "497c4c93-801e-41a6-ba8d-1c2f51710546" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.555s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.921056] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.921056] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.921056] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.921056] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.921619] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.923643] env[62914]: INFO nova.compute.manager [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Terminating instance [ 1184.925525] env[62914]: DEBUG nova.compute.manager [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1184.925726] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1184.926565] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6b6a11-566c-4d48-a86e-653dbd29993c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.933945] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1184.934173] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d8d2dc3-5ef8-4a0f-894f-d0cb328ea675 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.940370] env[62914]: DEBUG oslo_vmware.api [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1184.940370] env[62914]: value = "task-1353027" [ 1184.940370] env[62914]: _type = "Task" [ 1184.940370] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.947765] env[62914]: DEBUG oslo_vmware.api [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353027, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.450098] env[62914]: DEBUG oslo_vmware.api [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353027, 'name': PowerOffVM_Task, 'duration_secs': 0.19644} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.450402] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1185.450582] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1185.450852] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c5a5270-2740-499e-a0ea-9c33b0aa7e44 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.510252] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1185.510495] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1185.510688] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Deleting the datastore file [datastore1] 1005f4b0-e07a-4f32-9b3d-a2868a307bb9 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1185.510960] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a11d80b2-0ff7-4811-8f53-45740c2d6382 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.517646] env[62914]: DEBUG oslo_vmware.api [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for the task: (returnval){ [ 1185.517646] env[62914]: value = "task-1353029" [ 1185.517646] env[62914]: _type = "Task" [ 1185.517646] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.525295] env[62914]: DEBUG oslo_vmware.api [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353029, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.027424] env[62914]: DEBUG oslo_vmware.api [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Task: {'id': task-1353029, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132703} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.027800] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1186.027893] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1186.028127] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1186.028327] env[62914]: INFO nova.compute.manager [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1186.028571] env[62914]: DEBUG oslo.service.loopingcall [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1186.028771] env[62914]: DEBUG nova.compute.manager [-] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1186.028866] env[62914]: DEBUG nova.network.neutron [-] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1186.749390] env[62914]: DEBUG nova.compute.manager [req-ea186b89-9163-428f-93ef-8553291e4d14 req-91b5e78e-783a-4e4a-95ec-dec439b9821d service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Received event network-vif-deleted-10301fa4-7a51-4bd7-8496-6e01598e68dc {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1186.749610] env[62914]: INFO nova.compute.manager [req-ea186b89-9163-428f-93ef-8553291e4d14 req-91b5e78e-783a-4e4a-95ec-dec439b9821d service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Neutron deleted interface 10301fa4-7a51-4bd7-8496-6e01598e68dc; detaching it from the instance and deleting it from the info cache [ 1186.749793] env[62914]: DEBUG nova.network.neutron [req-ea186b89-9163-428f-93ef-8553291e4d14 req-91b5e78e-783a-4e4a-95ec-dec439b9821d service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.191473] env[62914]: DEBUG nova.network.neutron [-] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.252113] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-53860fb5-958b-445b-9a93-cee86593dd0a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.261548] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27faa751-5ea0-4b4f-9513-c84aef20ae55 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.285650] env[62914]: DEBUG nova.compute.manager [req-ea186b89-9163-428f-93ef-8553291e4d14 req-91b5e78e-783a-4e4a-95ec-dec439b9821d service nova] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Detach interface failed, port_id=10301fa4-7a51-4bd7-8496-6e01598e68dc, reason: Instance 1005f4b0-e07a-4f32-9b3d-a2868a307bb9 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1187.382531] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "0e461e81-ec47-40b6-9c09-87037f628750" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.382834] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "0e461e81-ec47-40b6-9c09-87037f628750" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.694852] env[62914]: INFO nova.compute.manager [-] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Took 1.67 seconds to deallocate network for instance. [ 1187.885226] env[62914]: DEBUG nova.compute.manager [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1188.201641] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.202019] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.202154] env[62914]: DEBUG nova.objects.instance [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lazy-loading 'resources' on Instance uuid 1005f4b0-e07a-4f32-9b3d-a2868a307bb9 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.403808] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.749148] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da5647a-b2f4-4890-a9f9-93bd210dae6b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.756777] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed66a77-5dea-4b48-b991-ae61da846de3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.784907] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293e1a0a-37ec-452b-9801-7bc89c3af5b2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.791484] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb044236-f034-4eb0-b90a-82e7200ed21f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.805151] env[62914]: DEBUG nova.compute.provider_tree [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1189.308298] env[62914]: DEBUG nova.scheduler.client.report [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1189.813768] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.612s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.817047] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.412s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.819048] env[62914]: INFO nova.compute.claims [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1189.829615] env[62914]: INFO nova.scheduler.client.report [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Deleted allocations for instance 1005f4b0-e07a-4f32-9b3d-a2868a307bb9 [ 1190.336200] env[62914]: DEBUG oslo_concurrency.lockutils [None req-eb624e4e-1ad0-48f3-acf8-23e3fbdc3d6f tempest-AttachVolumeTestJSON-1516225875 tempest-AttachVolumeTestJSON-1516225875-project-member] Lock "1005f4b0-e07a-4f32-9b3d-a2868a307bb9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.415s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.867563] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a784749-7a0c-4067-b03d-e762bd62c372 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.875506] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db8158e4-1e8c-49c5-9785-5a50669a9b41 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.905932] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-398ce4ca-9723-4fcb-9c8a-083717bfe99c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.913026] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e762f73-3ac6-4d07-a400-9a5eb48a5174 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.927384] env[62914]: DEBUG nova.compute.provider_tree [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1191.430426] env[62914]: DEBUG nova.scheduler.client.report [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1191.935636] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.119s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.936223] env[62914]: DEBUG nova.compute.manager [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1192.441631] env[62914]: DEBUG nova.compute.utils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1192.443054] env[62914]: DEBUG nova.compute.manager [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1192.443232] env[62914]: DEBUG nova.network.neutron [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1192.491754] env[62914]: DEBUG nova.policy [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7ab20b70ceb4f5bb700de50494399cc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbb45359291e4d699c21832d0265c1dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 1192.758066] env[62914]: DEBUG nova.network.neutron [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Successfully created port: 8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1192.950189] env[62914]: DEBUG nova.compute.manager [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1193.962292] env[62914]: DEBUG nova.compute.manager [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1193.988350] env[62914]: DEBUG nova.virt.hardware [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1193.988623] env[62914]: DEBUG nova.virt.hardware [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1193.988789] env[62914]: DEBUG nova.virt.hardware [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1193.988980] env[62914]: DEBUG nova.virt.hardware [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1193.989153] env[62914]: DEBUG nova.virt.hardware [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1193.989310] env[62914]: DEBUG nova.virt.hardware [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1193.989527] env[62914]: DEBUG nova.virt.hardware [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1193.989695] env[62914]: DEBUG nova.virt.hardware [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1193.989869] env[62914]: DEBUG nova.virt.hardware [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1193.990053] env[62914]: DEBUG nova.virt.hardware [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1193.990245] env[62914]: DEBUG nova.virt.hardware [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1193.991116] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a81a5d-0221-4f6d-9139-65876b860a21 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.998960] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51bf1814-9177-49e0-b9b5-3765cb7920ad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.139465] env[62914]: DEBUG nova.compute.manager [req-67ed922a-902c-4ab7-822f-866a1d6bf025 req-7d2b76f3-4b61-48e0-95b0-fa94bbb23c59 service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Received event network-vif-plugged-8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1194.139700] env[62914]: DEBUG oslo_concurrency.lockutils [req-67ed922a-902c-4ab7-822f-866a1d6bf025 req-7d2b76f3-4b61-48e0-95b0-fa94bbb23c59 service nova] Acquiring lock "0e461e81-ec47-40b6-9c09-87037f628750-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.139918] env[62914]: DEBUG oslo_concurrency.lockutils [req-67ed922a-902c-4ab7-822f-866a1d6bf025 req-7d2b76f3-4b61-48e0-95b0-fa94bbb23c59 service nova] Lock "0e461e81-ec47-40b6-9c09-87037f628750-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.140108] env[62914]: DEBUG oslo_concurrency.lockutils [req-67ed922a-902c-4ab7-822f-866a1d6bf025 req-7d2b76f3-4b61-48e0-95b0-fa94bbb23c59 service nova] Lock "0e461e81-ec47-40b6-9c09-87037f628750-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.140283] env[62914]: DEBUG nova.compute.manager [req-67ed922a-902c-4ab7-822f-866a1d6bf025 req-7d2b76f3-4b61-48e0-95b0-fa94bbb23c59 service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] No waiting events found dispatching network-vif-plugged-8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1194.140458] env[62914]: WARNING nova.compute.manager [req-67ed922a-902c-4ab7-822f-866a1d6bf025 req-7d2b76f3-4b61-48e0-95b0-fa94bbb23c59 service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Received unexpected event network-vif-plugged-8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5 for instance with vm_state building and task_state spawning. [ 1194.223746] env[62914]: DEBUG nova.network.neutron [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Successfully updated port: 8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1194.726207] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "refresh_cache-0e461e81-ec47-40b6-9c09-87037f628750" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1194.726357] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "refresh_cache-0e461e81-ec47-40b6-9c09-87037f628750" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1194.726506] env[62914]: DEBUG nova.network.neutron [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1195.258867] env[62914]: DEBUG nova.network.neutron [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1195.412256] env[62914]: DEBUG nova.network.neutron [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Updating instance_info_cache with network_info: [{"id": "8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5", "address": "fa:16:3e:4e:a5:cd", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f5bcb43-5f", "ovs_interfaceid": "8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.915114] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "refresh_cache-0e461e81-ec47-40b6-9c09-87037f628750" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1195.915266] env[62914]: DEBUG nova.compute.manager [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Instance network_info: |[{"id": "8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5", "address": "fa:16:3e:4e:a5:cd", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f5bcb43-5f", "ovs_interfaceid": "8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1195.915659] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:a5:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '504a3211-8288-4b87-a52a-590466d80a49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1195.923241] env[62914]: DEBUG oslo.service.loopingcall [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1195.923501] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1195.923724] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b178b075-1318-4da5-9b38-59384e0200e2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.944565] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1195.944565] env[62914]: value = "task-1353032" [ 1195.944565] env[62914]: _type = "Task" [ 1195.944565] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.952308] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1353032, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.165299] env[62914]: DEBUG nova.compute.manager [req-529875f5-8892-48e0-9839-05ca66b2c178 req-855c5ad1-78bd-481b-af9b-b3eefbaaa38a service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Received event network-changed-8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1196.165466] env[62914]: DEBUG nova.compute.manager [req-529875f5-8892-48e0-9839-05ca66b2c178 req-855c5ad1-78bd-481b-af9b-b3eefbaaa38a service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Refreshing instance network info cache due to event network-changed-8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1196.165693] env[62914]: DEBUG oslo_concurrency.lockutils [req-529875f5-8892-48e0-9839-05ca66b2c178 req-855c5ad1-78bd-481b-af9b-b3eefbaaa38a service nova] Acquiring lock "refresh_cache-0e461e81-ec47-40b6-9c09-87037f628750" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1196.165898] env[62914]: DEBUG oslo_concurrency.lockutils [req-529875f5-8892-48e0-9839-05ca66b2c178 req-855c5ad1-78bd-481b-af9b-b3eefbaaa38a service nova] Acquired lock "refresh_cache-0e461e81-ec47-40b6-9c09-87037f628750" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.166131] env[62914]: DEBUG nova.network.neutron [req-529875f5-8892-48e0-9839-05ca66b2c178 req-855c5ad1-78bd-481b-af9b-b3eefbaaa38a service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Refreshing network info cache for port 8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1196.454466] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1353032, 'name': CreateVM_Task, 'duration_secs': 0.274442} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.454881] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1196.455380] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1196.455537] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.455981] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1196.456247] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34c12036-f6cd-4e8e-a1b3-7e387e9c2842 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.460739] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1196.460739] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52b0267c-fd06-4f80-e634-cfbbdf0658b8" [ 1196.460739] env[62914]: _type = "Task" [ 1196.460739] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.468038] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52b0267c-fd06-4f80-e634-cfbbdf0658b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.862157] env[62914]: DEBUG nova.network.neutron [req-529875f5-8892-48e0-9839-05ca66b2c178 req-855c5ad1-78bd-481b-af9b-b3eefbaaa38a service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Updated VIF entry in instance network info cache for port 8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1196.862615] env[62914]: DEBUG nova.network.neutron [req-529875f5-8892-48e0-9839-05ca66b2c178 req-855c5ad1-78bd-481b-af9b-b3eefbaaa38a service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Updating instance_info_cache with network_info: [{"id": "8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5", "address": "fa:16:3e:4e:a5:cd", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f5bcb43-5f", "ovs_interfaceid": "8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1196.971112] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52b0267c-fd06-4f80-e634-cfbbdf0658b8, 'name': SearchDatastore_Task, 'duration_secs': 0.010422} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.971415] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1196.971684] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1196.971899] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1196.972068] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.972256] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1196.972514] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9356b4d2-c7d8-47a3-bc77-f0142d00d3df {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.980149] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1196.980340] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1196.980995] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52de21db-c09c-416c-af67-446cd81ab1b9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.985851] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1196.985851] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52916516-4a80-a489-57f3-5fe8f1858c44" [ 1196.985851] env[62914]: _type = "Task" [ 1196.985851] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.992642] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52916516-4a80-a489-57f3-5fe8f1858c44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.365867] env[62914]: DEBUG oslo_concurrency.lockutils [req-529875f5-8892-48e0-9839-05ca66b2c178 req-855c5ad1-78bd-481b-af9b-b3eefbaaa38a service nova] Releasing lock "refresh_cache-0e461e81-ec47-40b6-9c09-87037f628750" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1197.497012] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52916516-4a80-a489-57f3-5fe8f1858c44, 'name': SearchDatastore_Task, 'duration_secs': 0.007982} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.497774] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb623cb8-b267-41ed-93d0-b5477da7d584 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.502823] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1197.502823] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52befc75-c536-8348-971a-6942511a67f0" [ 1197.502823] env[62914]: _type = "Task" [ 1197.502823] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.510298] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52befc75-c536-8348-971a-6942511a67f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.013105] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52befc75-c536-8348-971a-6942511a67f0, 'name': SearchDatastore_Task, 'duration_secs': 0.009204} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.013371] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1198.013670] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 0e461e81-ec47-40b6-9c09-87037f628750/0e461e81-ec47-40b6-9c09-87037f628750.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1198.013929] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b44e1f94-c235-4b2e-b10f-df5cde930e52 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.019893] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1198.019893] env[62914]: value = "task-1353034" [ 1198.019893] env[62914]: _type = "Task" [ 1198.019893] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.026918] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353034, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.529824] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353034, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.420704} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.530207] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 0e461e81-ec47-40b6-9c09-87037f628750/0e461e81-ec47-40b6-9c09-87037f628750.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1198.530330] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1198.530555] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b30ea22b-98e8-4d79-b3e2-64af74cba08b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.537734] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1198.537734] env[62914]: value = "task-1353035" [ 1198.537734] env[62914]: _type = "Task" [ 1198.537734] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.544734] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353035, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.047061] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353035, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.052559} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.047319] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1199.048068] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d8fbfd9-14a2-402d-b1e1-474d505de26f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.068887] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 0e461e81-ec47-40b6-9c09-87037f628750/0e461e81-ec47-40b6-9c09-87037f628750.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1199.069131] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5e53cec-23e5-4ff9-808a-3ed6761a0054 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.087252] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1199.087252] env[62914]: value = "task-1353036" [ 1199.087252] env[62914]: _type = "Task" [ 1199.087252] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.094548] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353036, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.597335] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353036, 'name': ReconfigVM_Task, 'duration_secs': 0.250594} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.597672] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 0e461e81-ec47-40b6-9c09-87037f628750/0e461e81-ec47-40b6-9c09-87037f628750.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1199.598245] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d7e16b9-ed74-4394-952b-adef0078416a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.605947] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1199.605947] env[62914]: value = "task-1353037" [ 1199.605947] env[62914]: _type = "Task" [ 1199.605947] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.612970] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353037, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.115709] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353037, 'name': Rename_Task, 'duration_secs': 0.141053} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.115984] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1200.116246] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c7ee4b89-a3a8-4d78-9934-322c2c0aebf7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.122691] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1200.122691] env[62914]: value = "task-1353038" [ 1200.122691] env[62914]: _type = "Task" [ 1200.122691] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.130140] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353038, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.632421] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353038, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.132642] env[62914]: DEBUG oslo_vmware.api [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353038, 'name': PowerOnVM_Task, 'duration_secs': 0.523356} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.132881] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1201.133098] env[62914]: INFO nova.compute.manager [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Took 7.17 seconds to spawn the instance on the hypervisor. [ 1201.133285] env[62914]: DEBUG nova.compute.manager [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1201.134078] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86611f78-a0ed-4d55-bc21-9891d584a9d0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.650894] env[62914]: INFO nova.compute.manager [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Took 13.26 seconds to build instance. [ 1202.118542] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.118852] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.153132] env[62914]: DEBUG oslo_concurrency.lockutils [None req-f1245be3-fa0c-425c-81b9-dc57aad0fe29 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "0e461e81-ec47-40b6-9c09-87037f628750" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.770s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.624458] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.624682] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Starting heal instance info cache {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1202.708963] env[62914]: DEBUG nova.compute.manager [req-a95ee7ed-3cf7-46db-8301-ad339725e499 req-6abc708b-ba64-424b-8266-ecb199c6eae4 service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Received event network-changed-8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1202.708963] env[62914]: DEBUG nova.compute.manager [req-a95ee7ed-3cf7-46db-8301-ad339725e499 req-6abc708b-ba64-424b-8266-ecb199c6eae4 service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Refreshing instance network info cache due to event network-changed-8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1202.709487] env[62914]: DEBUG oslo_concurrency.lockutils [req-a95ee7ed-3cf7-46db-8301-ad339725e499 req-6abc708b-ba64-424b-8266-ecb199c6eae4 service nova] Acquiring lock "refresh_cache-0e461e81-ec47-40b6-9c09-87037f628750" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1202.709487] env[62914]: DEBUG oslo_concurrency.lockutils [req-a95ee7ed-3cf7-46db-8301-ad339725e499 req-6abc708b-ba64-424b-8266-ecb199c6eae4 service nova] Acquired lock "refresh_cache-0e461e81-ec47-40b6-9c09-87037f628750" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1202.709487] env[62914]: DEBUG nova.network.neutron [req-a95ee7ed-3cf7-46db-8301-ad339725e499 req-6abc708b-ba64-424b-8266-ecb199c6eae4 service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Refreshing network info cache for port 8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1203.127503] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Didn't find any instances for network info cache update. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1203.127726] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.127895] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.128392] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.128392] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.128392] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.128541] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.128624] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62914) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1203.128743] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager.update_available_resource {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.414464] env[62914]: DEBUG nova.network.neutron [req-a95ee7ed-3cf7-46db-8301-ad339725e499 req-6abc708b-ba64-424b-8266-ecb199c6eae4 service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Updated VIF entry in instance network info cache for port 8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1203.414903] env[62914]: DEBUG nova.network.neutron [req-a95ee7ed-3cf7-46db-8301-ad339725e499 req-6abc708b-ba64-424b-8266-ecb199c6eae4 service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Updating instance_info_cache with network_info: [{"id": "8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5", "address": "fa:16:3e:4e:a5:cd", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f5bcb43-5f", "ovs_interfaceid": "8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1203.631464] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.631650] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.631824] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.631979] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62914) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1203.632894] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499d3f08-c652-459c-ae20-156e3fd32e59 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.641162] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e1a900-e5c7-4239-a209-880960851f7a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.654539] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15fc8f6e-b8a5-4e21-a51e-e91acdc85b7e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.660600] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79e41fd-7303-4667-abcb-5326d6fdbcf5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.688312] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181117MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62914) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1203.688473] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.688663] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.917660] env[62914]: DEBUG oslo_concurrency.lockutils [req-a95ee7ed-3cf7-46db-8301-ad339725e499 req-6abc708b-ba64-424b-8266-ecb199c6eae4 service nova] Releasing lock "refresh_cache-0e461e81-ec47-40b6-9c09-87037f628750" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1204.715391] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance f1c5c877-9c29-4331-84f8-23b3c6b53725 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1204.715551] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 0e461e81-ec47-40b6-9c09-87037f628750 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1204.715736] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1204.715899] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1204.749846] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2a0117-2b06-41c9-8a5a-2dc95218ae76 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.757625] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5b1b9e-9842-46dd-a258-4cdf129a856a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.786567] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58efb733-3e17-451b-8955-4bc2ec5f897f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.793026] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ced70a-6111-4ba1-be25-fc14f13e4106 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.805210] env[62914]: DEBUG nova.compute.provider_tree [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1205.308711] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1205.813996] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1205.814232] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.126s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.492897] env[62914]: DEBUG oslo_concurrency.lockutils [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "0e461e81-ec47-40b6-9c09-87037f628750" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1239.493184] env[62914]: DEBUG oslo_concurrency.lockutils [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "0e461e81-ec47-40b6-9c09-87037f628750" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1239.996082] env[62914]: DEBUG nova.compute.utils [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1240.499258] env[62914]: DEBUG oslo_concurrency.lockutils [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "0e461e81-ec47-40b6-9c09-87037f628750" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1241.555501] env[62914]: DEBUG oslo_concurrency.lockutils [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "0e461e81-ec47-40b6-9c09-87037f628750" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1241.555884] env[62914]: DEBUG oslo_concurrency.lockutils [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "0e461e81-ec47-40b6-9c09-87037f628750" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1241.555988] env[62914]: INFO nova.compute.manager [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Attaching volume 6128d3da-447f-4a0f-a9af-906d31636820 to /dev/sdb [ 1241.585783] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc2198d-067d-4fea-a8b6-a8e19f804508 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.592895] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1514093a-217e-4181-8223-9b03e6fb5476 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.606662] env[62914]: DEBUG nova.virt.block_device [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Updating existing volume attachment record: 6cc0371b-4101-425e-b8b5-a159c24bb541 {{(pid=62914) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1246.149686] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Volume attach. Driver type: vmdk {{(pid=62914) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1246.149929] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288279', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'name': 'volume-6128d3da-447f-4a0f-a9af-906d31636820', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0e461e81-ec47-40b6-9c09-87037f628750', 'attached_at': '', 'detached_at': '', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'serial': '6128d3da-447f-4a0f-a9af-906d31636820'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1246.150828] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed32e0e9-57ce-4079-9193-ade0340e13f9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.166950] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557d2a4b-1539-42dd-a7f4-ab1f31038aa2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.190243] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] volume-6128d3da-447f-4a0f-a9af-906d31636820/volume-6128d3da-447f-4a0f-a9af-906d31636820.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1246.190477] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a974157-36c9-49e2-8d8c-91fe7887fae4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.207645] env[62914]: DEBUG oslo_vmware.api [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1246.207645] env[62914]: value = "task-1353043" [ 1246.207645] env[62914]: _type = "Task" [ 1246.207645] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.215670] env[62914]: DEBUG oslo_vmware.api [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353043, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.718881] env[62914]: DEBUG oslo_vmware.api [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353043, 'name': ReconfigVM_Task, 'duration_secs': 0.341714} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.719211] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Reconfigured VM instance instance-0000006a to attach disk [datastore1] volume-6128d3da-447f-4a0f-a9af-906d31636820/volume-6128d3da-447f-4a0f-a9af-906d31636820.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1246.723617] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1f53590-79bf-4995-a46d-8679ee4ce41b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.737513] env[62914]: DEBUG oslo_vmware.api [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1246.737513] env[62914]: value = "task-1353044" [ 1246.737513] env[62914]: _type = "Task" [ 1246.737513] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.744705] env[62914]: DEBUG oslo_vmware.api [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353044, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.246896] env[62914]: DEBUG oslo_vmware.api [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353044, 'name': ReconfigVM_Task, 'duration_secs': 0.127873} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.247238] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288279', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'name': 'volume-6128d3da-447f-4a0f-a9af-906d31636820', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0e461e81-ec47-40b6-9c09-87037f628750', 'attached_at': '', 'detached_at': '', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'serial': '6128d3da-447f-4a0f-a9af-906d31636820'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1248.283139] env[62914]: DEBUG nova.objects.instance [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lazy-loading 'flavor' on Instance uuid 0e461e81-ec47-40b6-9c09-87037f628750 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1248.787055] env[62914]: DEBUG oslo_concurrency.lockutils [None req-030445f4-bbe5-453d-a234-51b9dbcd7e05 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "0e461e81-ec47-40b6-9c09-87037f628750" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.231s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.122693] env[62914]: INFO nova.compute.manager [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Rebuilding instance [ 1249.163831] env[62914]: DEBUG nova.compute.manager [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1249.164728] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0842b5b3-2f03-4bc6-88f0-61d5bf0c48c0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.675809] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1249.676179] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc5f5c9c-d435-45e7-b407-9dc7dcf3b5da {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.683396] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1249.683396] env[62914]: value = "task-1353045" [ 1249.683396] env[62914]: _type = "Task" [ 1249.683396] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.692473] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353045, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.193119] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353045, 'name': PowerOffVM_Task, 'duration_secs': 0.201961} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.193396] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1250.246435] env[62914]: INFO nova.compute.manager [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Detaching volume 6128d3da-447f-4a0f-a9af-906d31636820 [ 1250.275531] env[62914]: INFO nova.virt.block_device [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Attempting to driver detach volume 6128d3da-447f-4a0f-a9af-906d31636820 from mountpoint /dev/sdb [ 1250.275771] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Volume detach. Driver type: vmdk {{(pid=62914) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1250.275966] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288279', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'name': 'volume-6128d3da-447f-4a0f-a9af-906d31636820', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0e461e81-ec47-40b6-9c09-87037f628750', 'attached_at': '', 'detached_at': '', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'serial': '6128d3da-447f-4a0f-a9af-906d31636820'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1250.276872] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0037aaa-a8c5-4f29-a626-8f0fe3dda90b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.298349] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5310f3a6-7946-4df7-a496-4425d9e2a478 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.304658] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee42795-fe90-4d63-8d81-3f3696482a1b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.324357] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b53af36-28a8-4d65-b48f-b952b8493da0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.338782] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] The volume has not been displaced from its original location: [datastore1] volume-6128d3da-447f-4a0f-a9af-906d31636820/volume-6128d3da-447f-4a0f-a9af-906d31636820.vmdk. No consolidation needed. {{(pid=62914) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1250.343904] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1250.344176] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b3bca69-620e-4ca2-96eb-6e11c78419f1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.360783] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1250.360783] env[62914]: value = "task-1353046" [ 1250.360783] env[62914]: _type = "Task" [ 1250.360783] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.367948] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353046, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.870677] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353046, 'name': ReconfigVM_Task, 'duration_secs': 0.173852} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.870978] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1250.875523] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52346f49-5fea-47d8-823c-3008cbe03804 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.889750] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1250.889750] env[62914]: value = "task-1353047" [ 1250.889750] env[62914]: _type = "Task" [ 1250.889750] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.897569] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353047, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.399586] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353047, 'name': ReconfigVM_Task, 'duration_secs': 0.138119} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.399995] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288279', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'name': 'volume-6128d3da-447f-4a0f-a9af-906d31636820', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0e461e81-ec47-40b6-9c09-87037f628750', 'attached_at': '', 'detached_at': '', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'serial': '6128d3da-447f-4a0f-a9af-906d31636820'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1252.447712] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1252.448077] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b938cc0d-921e-4b9f-948c-90cf602166a4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.455478] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1252.455478] env[62914]: value = "task-1353048" [ 1252.455478] env[62914]: _type = "Task" [ 1252.455478] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.462869] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353048, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.966321] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] VM already powered off {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1252.966585] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Volume detach. Driver type: vmdk {{(pid=62914) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1252.966833] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288279', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'name': 'volume-6128d3da-447f-4a0f-a9af-906d31636820', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0e461e81-ec47-40b6-9c09-87037f628750', 'attached_at': '', 'detached_at': '', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'serial': '6128d3da-447f-4a0f-a9af-906d31636820'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1252.967606] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658886e0-c2af-47bc-b86d-02d50e5482db {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.986397] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba5f2a5b-4d7b-44cd-b55a-816d504554cb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.992036] env[62914]: WARNING nova.virt.vmwareapi.driver [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1252.992313] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1252.993010] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3cdf722-911f-4d6a-80b9-d2309b2acf01 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.998914] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1252.999163] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-631cd5d6-40db-4aab-9a84-67ae5821cd9c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.066838] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1253.067121] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1253.067298] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleting the datastore file [datastore1] 0e461e81-ec47-40b6-9c09-87037f628750 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1253.067558] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-298208c6-62e8-4893-96bf-e38eddd4f51f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.073417] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1253.073417] env[62914]: value = "task-1353050" [ 1253.073417] env[62914]: _type = "Task" [ 1253.073417] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.081519] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353050, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.583127] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353050, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137694} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.583503] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1253.583661] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1253.583766] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1254.089348] env[62914]: INFO nova.virt.block_device [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Booting with volume 6128d3da-447f-4a0f-a9af-906d31636820 at /dev/sdb [ 1254.118890] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5a9d0cba-41de-476d-9402-55622c624e35 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.128642] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82105475-c0c1-4f06-9f18-235503ae5a41 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.150036] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c9928545-85f9-4393-8d76-6ba1db191dad {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.157027] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb79a655-4b19-41a9-8a47-9a5399d33788 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.178212] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887462e9-c95d-4b9a-9ad5-070b68abc820 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.184024] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc1cc23-74d0-4eb7-abdd-168ce74cd8c3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.196683] env[62914]: DEBUG nova.virt.block_device [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Updating existing volume attachment record: 8a1f67e3-91c1-47d0-be20-42cb4cddbaa9 {{(pid=62914) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1256.299187] env[62914]: DEBUG nova.virt.hardware [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1256.299454] env[62914]: DEBUG nova.virt.hardware [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1256.299617] env[62914]: DEBUG nova.virt.hardware [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1256.299802] env[62914]: DEBUG nova.virt.hardware [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1256.299952] env[62914]: DEBUG nova.virt.hardware [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1256.300121] env[62914]: DEBUG nova.virt.hardware [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1256.300335] env[62914]: DEBUG nova.virt.hardware [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1256.300495] env[62914]: DEBUG nova.virt.hardware [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1256.300660] env[62914]: DEBUG nova.virt.hardware [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1256.300824] env[62914]: DEBUG nova.virt.hardware [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1256.300996] env[62914]: DEBUG nova.virt.hardware [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1256.301863] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af32dc12-cee0-4009-819b-c4d0c4882086 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.309749] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83c1666-f3cc-4572-bc5a-5df4d24eee56 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.322909] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:a5:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '504a3211-8288-4b87-a52a-590466d80a49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1256.330087] env[62914]: DEBUG oslo.service.loopingcall [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1256.330330] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1256.330535] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0047626a-01a8-4db1-b5de-a339b9acfc32 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.348445] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1256.348445] env[62914]: value = "task-1353051" [ 1256.348445] env[62914]: _type = "Task" [ 1256.348445] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.355298] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1353051, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.857841] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1353051, 'name': CreateVM_Task, 'duration_secs': 0.300537} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.858021] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1256.858699] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.858874] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.859232] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1256.859484] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-001a3acf-3481-4f02-bdfd-dc609a29d9e8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.863716] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1256.863716] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5244e2f5-bdfc-1e9f-4f1e-c94888cd135f" [ 1256.863716] env[62914]: _type = "Task" [ 1256.863716] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.870768] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5244e2f5-bdfc-1e9f-4f1e-c94888cd135f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.374104] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5244e2f5-bdfc-1e9f-4f1e-c94888cd135f, 'name': SearchDatastore_Task, 'duration_secs': 0.01139} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.374629] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1257.374673] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1257.374867] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1257.375064] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1257.375266] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1257.375523] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f710c70b-68fa-4725-b0ff-8968cf523d32 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.383055] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1257.383237] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1257.383879] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12fc083e-c15f-4b50-ad0e-6f3a855c1107 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.388619] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1257.388619] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]5230b3d4-e4d6-3141-d1fc-cc8da1081462" [ 1257.388619] env[62914]: _type = "Task" [ 1257.388619] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.395475] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5230b3d4-e4d6-3141-d1fc-cc8da1081462, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.899230] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]5230b3d4-e4d6-3141-d1fc-cc8da1081462, 'name': SearchDatastore_Task, 'duration_secs': 0.007324} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.899963] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ca83ad0-9e1f-4334-8808-a50ab2e5cc13 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.905635] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1257.905635] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52828e05-1b91-5cf0-d7db-318ec03d70e4" [ 1257.905635] env[62914]: _type = "Task" [ 1257.905635] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.912756] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52828e05-1b91-5cf0-d7db-318ec03d70e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.416049] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52828e05-1b91-5cf0-d7db-318ec03d70e4, 'name': SearchDatastore_Task, 'duration_secs': 0.008879} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.416485] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1258.416591] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 0e461e81-ec47-40b6-9c09-87037f628750/0e461e81-ec47-40b6-9c09-87037f628750.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1258.416849] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce5c5c0f-c2ee-4d91-a47d-dfa45d34ae54 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.423479] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1258.423479] env[62914]: value = "task-1353052" [ 1258.423479] env[62914]: _type = "Task" [ 1258.423479] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.430345] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353052, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.933822] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353052, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457026} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.934227] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 0e461e81-ec47-40b6-9c09-87037f628750/0e461e81-ec47-40b6-9c09-87037f628750.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1258.934456] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1258.934749] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f195fb2-10ea-42f1-9610-191cbacd131f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.940297] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1258.940297] env[62914]: value = "task-1353053" [ 1258.940297] env[62914]: _type = "Task" [ 1258.940297] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.948848] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353053, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.450329] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353053, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055789} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.450683] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1259.451359] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c067037-4870-4ac9-b801-9ed87d869132 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.472189] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 0e461e81-ec47-40b6-9c09-87037f628750/0e461e81-ec47-40b6-9c09-87037f628750.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1259.472421] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c06b3159-e8a3-46f1-8cfa-387ded56a9b1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.491261] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1259.491261] env[62914]: value = "task-1353054" [ 1259.491261] env[62914]: _type = "Task" [ 1259.491261] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.500626] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353054, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.000842] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353054, 'name': ReconfigVM_Task, 'duration_secs': 0.268514} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.001090] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 0e461e81-ec47-40b6-9c09-87037f628750/0e461e81-ec47-40b6-9c09-87037f628750.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1260.002472] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'disk_bus': None, 'device_name': '/dev/sda', 'guest_format': None, 'size': 0, 'encryption_secret_uuid': None, 'encryption_format': None, 'encryption_options': None, 'encrypted': False, 'device_type': 'disk', 'boot_index': 0, 'image_id': '6c21e11e-da68-48a7-9929-737212ce98cf'}], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': False, 'attachment_id': '8a1f67e3-91c1-47d0-be20-42cb4cddbaa9', 'disk_bus': None, 'guest_format': None, 'mount_device': '/dev/sdb', 'device_type': None, 'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288279', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'name': 'volume-6128d3da-447f-4a0f-a9af-906d31636820', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0e461e81-ec47-40b6-9c09-87037f628750', 'attached_at': '', 'detached_at': '', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'serial': '6128d3da-447f-4a0f-a9af-906d31636820'}, 'volume_type': None}], 'swap': None} {{(pid=62914) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1260.002681] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Volume attach. Driver type: vmdk {{(pid=62914) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1260.002879] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288279', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'name': 'volume-6128d3da-447f-4a0f-a9af-906d31636820', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0e461e81-ec47-40b6-9c09-87037f628750', 'attached_at': '', 'detached_at': '', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'serial': '6128d3da-447f-4a0f-a9af-906d31636820'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1260.003652] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2e6fa0-0d51-4fe7-b01d-e6ff4f83a32b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.018404] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e67688-ddac-48db-be78-0a4f4da78c95 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.041274] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] volume-6128d3da-447f-4a0f-a9af-906d31636820/volume-6128d3da-447f-4a0f-a9af-906d31636820.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1260.041421] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b3fb5a7-4499-45e5-a6a0-ba7214311991 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.058306] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1260.058306] env[62914]: value = "task-1353055" [ 1260.058306] env[62914]: _type = "Task" [ 1260.058306] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.065172] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353055, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.568735] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353055, 'name': ReconfigVM_Task, 'duration_secs': 0.278757} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.569174] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Reconfigured VM instance instance-0000006a to attach disk [datastore1] volume-6128d3da-447f-4a0f-a9af-906d31636820/volume-6128d3da-447f-4a0f-a9af-906d31636820.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1260.573642] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f73e74fb-7c35-47c9-8538-bcf14dd67810 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.587736] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1260.587736] env[62914]: value = "task-1353056" [ 1260.587736] env[62914]: _type = "Task" [ 1260.587736] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.595219] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353056, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.098102] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353056, 'name': ReconfigVM_Task, 'duration_secs': 0.131906} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.098386] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288279', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'name': 'volume-6128d3da-447f-4a0f-a9af-906d31636820', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0e461e81-ec47-40b6-9c09-87037f628750', 'attached_at': '', 'detached_at': '', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'serial': '6128d3da-447f-4a0f-a9af-906d31636820'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1261.098969] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-70b5f63d-a916-437f-9799-00d019b5b4a2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.105200] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1261.105200] env[62914]: value = "task-1353057" [ 1261.105200] env[62914]: _type = "Task" [ 1261.105200] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.112177] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353057, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.615460] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353057, 'name': Rename_Task, 'duration_secs': 0.136054} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.615460] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1261.617356] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-25aab260-16bf-49d0-9d54-b3b8c26cd870 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.622218] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1261.622218] env[62914]: value = "task-1353058" [ 1261.622218] env[62914]: _type = "Task" [ 1261.622218] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.630549] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353058, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.135436] env[62914]: DEBUG oslo_vmware.api [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353058, 'name': PowerOnVM_Task, 'duration_secs': 0.433544} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.135741] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1262.136047] env[62914]: DEBUG nova.compute.manager [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1262.137102] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0435a19e-b99a-4ef2-9ac2-1cdf7d063905 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.653682] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.653682] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.653890] env[62914]: DEBUG nova.objects.instance [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62914) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1263.662179] env[62914]: DEBUG oslo_concurrency.lockutils [None req-12912c31-0da0-4b97-87c9-011b0b051ac2 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.008s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1264.062716] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "0e461e81-ec47-40b6-9c09-87037f628750" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1264.062978] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "0e461e81-ec47-40b6-9c09-87037f628750" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1264.566748] env[62914]: INFO nova.compute.manager [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Detaching volume 6128d3da-447f-4a0f-a9af-906d31636820 [ 1264.597844] env[62914]: INFO nova.virt.block_device [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Attempting to driver detach volume 6128d3da-447f-4a0f-a9af-906d31636820 from mountpoint /dev/sdb [ 1264.598135] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Volume detach. Driver type: vmdk {{(pid=62914) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1264.598349] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288279', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'name': 'volume-6128d3da-447f-4a0f-a9af-906d31636820', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0e461e81-ec47-40b6-9c09-87037f628750', 'attached_at': '', 'detached_at': '', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'serial': '6128d3da-447f-4a0f-a9af-906d31636820'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1264.599231] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5501cdba-0b0f-4211-93c0-9165458a93d6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.620692] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47390d21-1ec0-4388-bf3e-a09a665a4793 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.627521] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3944b9e8-7fc1-4395-a0ef-cb2c102cfe6b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.648358] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b427750b-b14d-4256-9d4b-b3e528027895 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.662507] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] The volume has not been displaced from its original location: [datastore1] volume-6128d3da-447f-4a0f-a9af-906d31636820/volume-6128d3da-447f-4a0f-a9af-906d31636820.vmdk. No consolidation needed. {{(pid=62914) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1264.667715] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1264.668047] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-153783e6-3590-4075-b8e1-d16065dba431 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.685229] env[62914]: DEBUG oslo_vmware.api [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1264.685229] env[62914]: value = "task-1353059" [ 1264.685229] env[62914]: _type = "Task" [ 1264.685229] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.692669] env[62914]: DEBUG oslo_vmware.api [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353059, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.195179] env[62914]: DEBUG oslo_vmware.api [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353059, 'name': ReconfigVM_Task, 'duration_secs': 0.216615} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.195477] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1265.199976] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc05b5ed-22ee-4bdd-a7d5-743cbbd97064 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.214678] env[62914]: DEBUG oslo_vmware.api [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1265.214678] env[62914]: value = "task-1353060" [ 1265.214678] env[62914]: _type = "Task" [ 1265.214678] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.222516] env[62914]: DEBUG oslo_vmware.api [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353060, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.724404] env[62914]: DEBUG oslo_vmware.api [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353060, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.815118] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.815350] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1265.815496] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Starting heal instance info cache {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1265.815613] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Rebuilding the list of instances to heal {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1266.225963] env[62914]: DEBUG oslo_vmware.api [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353060, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.349186] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1266.349338] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquired lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1266.349486] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Forcefully refreshing network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1266.349633] env[62914]: DEBUG nova.objects.instance [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lazy-loading 'info_cache' on Instance uuid f1c5c877-9c29-4331-84f8-23b3c6b53725 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1266.726599] env[62914]: DEBUG oslo_vmware.api [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353060, 'name': ReconfigVM_Task, 'duration_secs': 1.13084} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.726932] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288279', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'name': 'volume-6128d3da-447f-4a0f-a9af-906d31636820', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0e461e81-ec47-40b6-9c09-87037f628750', 'attached_at': '', 'detached_at': '', 'volume_id': '6128d3da-447f-4a0f-a9af-906d31636820', 'serial': '6128d3da-447f-4a0f-a9af-906d31636820'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1267.266261] env[62914]: DEBUG nova.objects.instance [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lazy-loading 'flavor' on Instance uuid 0e461e81-ec47-40b6-9c09-87037f628750 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1268.067391] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updating instance_info_cache with network_info: [{"id": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "address": "fa:16:3e:fb:87:48", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4f0d675-9d", "ovs_interfaceid": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1268.274340] env[62914]: DEBUG oslo_concurrency.lockutils [None req-6c437495-8f9e-441c-9c38-e1d78865286d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "0e461e81-ec47-40b6-9c09-87037f628750" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.211s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.570148] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Releasing lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1268.570311] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updated the network info_cache for instance {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1268.570499] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1268.570654] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1268.570796] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1268.570938] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1268.571127] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1268.571307] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1268.571436] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62914) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1268.571582] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager.update_available_resource {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1269.075357] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.075709] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.075756] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1269.075898] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62914) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1269.076857] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6219d587-1b80-4348-88c2-df36da78451b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.085441] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60fb8ccc-f2b7-4eaf-9631-d5757e7c277e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.098794] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-376b0c99-7598-4cbf-bb0b-c70506ac80a6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.104755] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bae745a-3652-43a0-a42d-78082cbb7026 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.132397] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181142MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62914) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1269.132541] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.132732] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.327444] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "0e461e81-ec47-40b6-9c09-87037f628750" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.327715] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "0e461e81-ec47-40b6-9c09-87037f628750" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.327923] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "0e461e81-ec47-40b6-9c09-87037f628750-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.328125] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "0e461e81-ec47-40b6-9c09-87037f628750-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.328337] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "0e461e81-ec47-40b6-9c09-87037f628750-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1269.330464] env[62914]: INFO nova.compute.manager [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Terminating instance [ 1269.333670] env[62914]: DEBUG nova.compute.manager [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1269.333863] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1269.334725] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d5e9cb-dd20-47d0-abb6-9c86db9bfba0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.342269] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1269.342459] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eeccd203-c7e8-47aa-b5da-186c08903d40 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.348291] env[62914]: DEBUG oslo_vmware.api [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1269.348291] env[62914]: value = "task-1353061" [ 1269.348291] env[62914]: _type = "Task" [ 1269.348291] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.355529] env[62914]: DEBUG oslo_vmware.api [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353061, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.858300] env[62914]: DEBUG oslo_vmware.api [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353061, 'name': PowerOffVM_Task, 'duration_secs': 0.247908} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.858594] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1269.858766] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1269.859024] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d23fb2b6-937e-4d10-b0a4-bd7b58f412db {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.160091] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance f1c5c877-9c29-4331-84f8-23b3c6b53725 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1270.160333] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 0e461e81-ec47-40b6-9c09-87037f628750 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1270.160436] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1270.160576] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1270.171858] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1270.172108] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1270.172898] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleting the datastore file [datastore1] 0e461e81-ec47-40b6-9c09-87037f628750 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1270.172898] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b737671-b5a5-4b7d-9241-df2e60118413 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.180061] env[62914]: DEBUG oslo_vmware.api [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1270.180061] env[62914]: value = "task-1353063" [ 1270.180061] env[62914]: _type = "Task" [ 1270.180061] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.187945] env[62914]: DEBUG oslo_vmware.api [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353063, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.200293] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb316697-061b-4219-8728-b794a68594e9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.206533] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56e0991-ac19-418f-917f-8032544f2b28 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.234745] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da21f189-fcb4-48c3-9918-58da7b7ac36c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.241206] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f4ae4a-5b1f-45d4-94b8-2bf0089bd446 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.253644] env[62914]: DEBUG nova.compute.provider_tree [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1270.690620] env[62914]: DEBUG oslo_vmware.api [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353063, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124275} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.690887] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1270.691066] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1270.691278] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1270.691464] env[62914]: INFO nova.compute.manager [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Took 1.36 seconds to destroy the instance on the hypervisor. [ 1270.691704] env[62914]: DEBUG oslo.service.loopingcall [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1270.691906] env[62914]: DEBUG nova.compute.manager [-] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1270.691990] env[62914]: DEBUG nova.network.neutron [-] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1270.756438] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1271.260618] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1271.260889] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.128s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.280903] env[62914]: DEBUG nova.compute.manager [req-a7f86000-12a3-4bae-9ab0-dd03971640d0 req-c2573909-8b1c-4e5b-bd08-724426fc8fe6 service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Received event network-vif-deleted-8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1271.280903] env[62914]: INFO nova.compute.manager [req-a7f86000-12a3-4bae-9ab0-dd03971640d0 req-c2573909-8b1c-4e5b-bd08-724426fc8fe6 service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Neutron deleted interface 8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5; detaching it from the instance and deleting it from the info cache [ 1271.280903] env[62914]: DEBUG nova.network.neutron [req-a7f86000-12a3-4bae-9ab0-dd03971640d0 req-c2573909-8b1c-4e5b-bd08-724426fc8fe6 service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1271.758999] env[62914]: DEBUG nova.network.neutron [-] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1271.783914] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e4cffc7-adea-4483-8a8b-8928af5596c0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.795258] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee80c43-fff6-4bbf-ad69-0aeffd4e6c90 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.818205] env[62914]: DEBUG nova.compute.manager [req-a7f86000-12a3-4bae-9ab0-dd03971640d0 req-c2573909-8b1c-4e5b-bd08-724426fc8fe6 service nova] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Detach interface failed, port_id=8f5bcb43-5f2b-4f84-ac82-97eb89fd3cf5, reason: Instance 0e461e81-ec47-40b6-9c09-87037f628750 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1272.261905] env[62914]: INFO nova.compute.manager [-] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Took 1.57 seconds to deallocate network for instance. [ 1272.769036] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1272.769323] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1272.769574] env[62914]: DEBUG nova.objects.instance [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lazy-loading 'resources' on Instance uuid 0e461e81-ec47-40b6-9c09-87037f628750 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1273.315158] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43964c13-f032-4439-af05-90ddb12475fc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.322654] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e7c9b1-d6fe-489a-8adc-be408ea65869 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.352415] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325dcb09-1f93-4b26-896c-481dcd0d660b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.359061] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c018aa3-79ae-424e-ac75-27f0813b1c1c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.371502] env[62914]: DEBUG nova.compute.provider_tree [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1273.874692] env[62914]: DEBUG nova.scheduler.client.report [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1274.379326] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.610s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1274.401028] env[62914]: INFO nova.scheduler.client.report [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleted allocations for instance 0e461e81-ec47-40b6-9c09-87037f628750 [ 1274.909059] env[62914]: DEBUG oslo_concurrency.lockutils [None req-4f34d028-a9c9-4c68-a476-92f08119fd0d tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "0e461e81-ec47-40b6-9c09-87037f628750" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.581s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1276.885263] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1276.885592] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1277.388017] env[62914]: DEBUG nova.compute.manager [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1277.911116] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1277.911385] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1277.912832] env[62914]: INFO nova.compute.claims [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1278.956828] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eaa328e-c5f2-43b3-ae6e-e290b74ef8dd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.964304] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a606215-d007-42d5-a138-c1a371fc5713 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.992517] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25737fbd-3ecc-4c65-9a7e-c4095162265f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.998895] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52bd4f2f-0468-4b08-8b15-b3bbe5507661 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.011207] env[62914]: DEBUG nova.compute.provider_tree [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1279.513886] env[62914]: DEBUG nova.scheduler.client.report [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1280.019489] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.108s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1280.020069] env[62914]: DEBUG nova.compute.manager [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1280.525693] env[62914]: DEBUG nova.compute.utils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1280.528406] env[62914]: DEBUG nova.compute.manager [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1280.528574] env[62914]: DEBUG nova.network.neutron [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1280.574276] env[62914]: DEBUG nova.policy [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7ab20b70ceb4f5bb700de50494399cc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbb45359291e4d699c21832d0265c1dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 1280.818634] env[62914]: DEBUG nova.network.neutron [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Successfully created port: 4542e979-2f0f-41ec-ab26-d6dbf11ec534 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1281.031960] env[62914]: DEBUG nova.compute.manager [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1282.042183] env[62914]: DEBUG nova.compute.manager [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1282.067334] env[62914]: DEBUG nova.virt.hardware [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T12:06:58Z,direct_url=,disk_format='vmdk',id=6c21e11e-da68-48a7-9929-737212ce98cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1380766398394232ae49fac4a30d56f3',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T12:06:58Z,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1282.067611] env[62914]: DEBUG nova.virt.hardware [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1282.067773] env[62914]: DEBUG nova.virt.hardware [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1282.067955] env[62914]: DEBUG nova.virt.hardware [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1282.068117] env[62914]: DEBUG nova.virt.hardware [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1282.068268] env[62914]: DEBUG nova.virt.hardware [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1282.068478] env[62914]: DEBUG nova.virt.hardware [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1282.068635] env[62914]: DEBUG nova.virt.hardware [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1282.068802] env[62914]: DEBUG nova.virt.hardware [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1282.068963] env[62914]: DEBUG nova.virt.hardware [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1282.069156] env[62914]: DEBUG nova.virt.hardware [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1282.069994] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de58f8a9-3acf-423d-9541-977578413e48 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.077703] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dec90ee-0178-496b-8491-f77da42c478d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.196652] env[62914]: DEBUG nova.compute.manager [req-75768874-b2e6-4911-a0fc-b401d04ff1e8 req-a550ac13-a1b7-45e1-ba24-5f9b2e847e5d service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Received event network-vif-plugged-4542e979-2f0f-41ec-ab26-d6dbf11ec534 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1282.196883] env[62914]: DEBUG oslo_concurrency.lockutils [req-75768874-b2e6-4911-a0fc-b401d04ff1e8 req-a550ac13-a1b7-45e1-ba24-5f9b2e847e5d service nova] Acquiring lock "4ee2aef3-a5ca-4049-88d5-0b2b0817bce7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1282.197120] env[62914]: DEBUG oslo_concurrency.lockutils [req-75768874-b2e6-4911-a0fc-b401d04ff1e8 req-a550ac13-a1b7-45e1-ba24-5f9b2e847e5d service nova] Lock "4ee2aef3-a5ca-4049-88d5-0b2b0817bce7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1282.197380] env[62914]: DEBUG oslo_concurrency.lockutils [req-75768874-b2e6-4911-a0fc-b401d04ff1e8 req-a550ac13-a1b7-45e1-ba24-5f9b2e847e5d service nova] Lock "4ee2aef3-a5ca-4049-88d5-0b2b0817bce7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1282.197506] env[62914]: DEBUG nova.compute.manager [req-75768874-b2e6-4911-a0fc-b401d04ff1e8 req-a550ac13-a1b7-45e1-ba24-5f9b2e847e5d service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] No waiting events found dispatching network-vif-plugged-4542e979-2f0f-41ec-ab26-d6dbf11ec534 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1282.197628] env[62914]: WARNING nova.compute.manager [req-75768874-b2e6-4911-a0fc-b401d04ff1e8 req-a550ac13-a1b7-45e1-ba24-5f9b2e847e5d service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Received unexpected event network-vif-plugged-4542e979-2f0f-41ec-ab26-d6dbf11ec534 for instance with vm_state building and task_state spawning. [ 1282.275454] env[62914]: DEBUG nova.network.neutron [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Successfully updated port: 4542e979-2f0f-41ec-ab26-d6dbf11ec534 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1282.777560] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "refresh_cache-4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1282.777806] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "refresh_cache-4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1282.777875] env[62914]: DEBUG nova.network.neutron [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1283.309239] env[62914]: DEBUG nova.network.neutron [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1283.429480] env[62914]: DEBUG nova.network.neutron [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Updating instance_info_cache with network_info: [{"id": "4542e979-2f0f-41ec-ab26-d6dbf11ec534", "address": "fa:16:3e:c8:6d:06", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4542e979-2f", "ovs_interfaceid": "4542e979-2f0f-41ec-ab26-d6dbf11ec534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1283.931808] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "refresh_cache-4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1283.932185] env[62914]: DEBUG nova.compute.manager [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Instance network_info: |[{"id": "4542e979-2f0f-41ec-ab26-d6dbf11ec534", "address": "fa:16:3e:c8:6d:06", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4542e979-2f", "ovs_interfaceid": "4542e979-2f0f-41ec-ab26-d6dbf11ec534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1283.932699] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:6d:06', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '504a3211-8288-4b87-a52a-590466d80a49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4542e979-2f0f-41ec-ab26-d6dbf11ec534', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1283.940179] env[62914]: DEBUG oslo.service.loopingcall [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1283.940401] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1283.940630] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-30f581da-575f-4152-92b4-fd84155a8829 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.960596] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1283.960596] env[62914]: value = "task-1353064" [ 1283.960596] env[62914]: _type = "Task" [ 1283.960596] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.968107] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1353064, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.229520] env[62914]: DEBUG nova.compute.manager [req-9e997ade-e054-4722-954a-7b4132c8f514 req-665b7c53-0c26-4ba2-9378-10909cbbda07 service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Received event network-changed-4542e979-2f0f-41ec-ab26-d6dbf11ec534 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1284.229730] env[62914]: DEBUG nova.compute.manager [req-9e997ade-e054-4722-954a-7b4132c8f514 req-665b7c53-0c26-4ba2-9378-10909cbbda07 service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Refreshing instance network info cache due to event network-changed-4542e979-2f0f-41ec-ab26-d6dbf11ec534. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1284.229954] env[62914]: DEBUG oslo_concurrency.lockutils [req-9e997ade-e054-4722-954a-7b4132c8f514 req-665b7c53-0c26-4ba2-9378-10909cbbda07 service nova] Acquiring lock "refresh_cache-4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1284.230198] env[62914]: DEBUG oslo_concurrency.lockutils [req-9e997ade-e054-4722-954a-7b4132c8f514 req-665b7c53-0c26-4ba2-9378-10909cbbda07 service nova] Acquired lock "refresh_cache-4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1284.230390] env[62914]: DEBUG nova.network.neutron [req-9e997ade-e054-4722-954a-7b4132c8f514 req-665b7c53-0c26-4ba2-9378-10909cbbda07 service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Refreshing network info cache for port 4542e979-2f0f-41ec-ab26-d6dbf11ec534 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1284.471989] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1353064, 'name': CreateVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.932081] env[62914]: DEBUG nova.network.neutron [req-9e997ade-e054-4722-954a-7b4132c8f514 req-665b7c53-0c26-4ba2-9378-10909cbbda07 service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Updated VIF entry in instance network info cache for port 4542e979-2f0f-41ec-ab26-d6dbf11ec534. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1284.932470] env[62914]: DEBUG nova.network.neutron [req-9e997ade-e054-4722-954a-7b4132c8f514 req-665b7c53-0c26-4ba2-9378-10909cbbda07 service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Updating instance_info_cache with network_info: [{"id": "4542e979-2f0f-41ec-ab26-d6dbf11ec534", "address": "fa:16:3e:c8:6d:06", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4542e979-2f", "ovs_interfaceid": "4542e979-2f0f-41ec-ab26-d6dbf11ec534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1284.970798] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1353064, 'name': CreateVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.435696] env[62914]: DEBUG oslo_concurrency.lockutils [req-9e997ade-e054-4722-954a-7b4132c8f514 req-665b7c53-0c26-4ba2-9378-10909cbbda07 service nova] Releasing lock "refresh_cache-4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1285.471330] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1353064, 'name': CreateVM_Task, 'duration_secs': 1.296312} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.471510] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1285.472142] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1285.472439] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1285.472620] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1285.472855] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b424edef-7e75-42c7-a9ed-d9d5e363e27f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.476849] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1285.476849] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]52812b45-f2eb-7f0d-2e85-1e40211da08a" [ 1285.476849] env[62914]: _type = "Task" [ 1285.476849] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.483604] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52812b45-f2eb-7f0d-2e85-1e40211da08a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.987267] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]52812b45-f2eb-7f0d-2e85-1e40211da08a, 'name': SearchDatastore_Task, 'duration_secs': 0.009733} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.987598] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1285.987840] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Processing image 6c21e11e-da68-48a7-9929-737212ce98cf {{(pid=62914) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1285.988091] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1285.988250] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1285.988434] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1285.988700] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ccf380a4-29ae-490c-8bd1-117e02a404fe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.996342] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62914) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1285.996533] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62914) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1285.997216] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58bcd42a-0878-42e2-addb-8c41e327517f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.001867] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1286.001867] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]521da57b-9c88-dce8-9266-58780406b8f2" [ 1286.001867] env[62914]: _type = "Task" [ 1286.001867] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.009285] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]521da57b-9c88-dce8-9266-58780406b8f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.512059] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]521da57b-9c88-dce8-9266-58780406b8f2, 'name': SearchDatastore_Task, 'duration_secs': 0.008131} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.512813] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02629c87-9fed-4796-8709-58dd02c01148 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.517564] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1286.517564] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]529eb8c8-d2c9-00cd-66df-03d55c46126f" [ 1286.517564] env[62914]: _type = "Task" [ 1286.517564] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.524433] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]529eb8c8-d2c9-00cd-66df-03d55c46126f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.028061] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]529eb8c8-d2c9-00cd-66df-03d55c46126f, 'name': SearchDatastore_Task, 'duration_secs': 0.00839} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.028332] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1287.028600] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7/4ee2aef3-a5ca-4049-88d5-0b2b0817bce7.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1287.028854] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d79cfb39-77dc-4a21-9595-da44ffc4c688 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.035349] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1287.035349] env[62914]: value = "task-1353065" [ 1287.035349] env[62914]: _type = "Task" [ 1287.035349] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.042292] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353065, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.544813] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353065, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.046641] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353065, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.552999] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353065, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.41706} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.553456] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c21e11e-da68-48a7-9929-737212ce98cf/6c21e11e-da68-48a7-9929-737212ce98cf.vmdk to [datastore1] 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7/4ee2aef3-a5ca-4049-88d5-0b2b0817bce7.vmdk {{(pid=62914) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1288.553768] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Extending root virtual disk to 1048576 {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1288.554146] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-373e8b0f-89db-4fd5-a35d-eea986d11e49 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.562717] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1288.562717] env[62914]: value = "task-1353066" [ 1288.562717] env[62914]: _type = "Task" [ 1288.562717] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.573548] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353066, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.072600] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353066, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.38593} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.072882] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Extended root virtual disk {{(pid=62914) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1289.073688] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72fe1266-3686-408e-954b-58120396630f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.095813] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7/4ee2aef3-a5ca-4049-88d5-0b2b0817bce7.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1289.096137] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8db2588f-e88a-46cc-b479-174b6f45da1b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.115590] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1289.115590] env[62914]: value = "task-1353067" [ 1289.115590] env[62914]: _type = "Task" [ 1289.115590] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.123691] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353067, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.625421] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353067, 'name': ReconfigVM_Task, 'duration_secs': 0.267747} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.625834] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7/4ee2aef3-a5ca-4049-88d5-0b2b0817bce7.vmdk or device None with type sparse {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1289.626296] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-282d8abe-a1e0-4c2e-a6b1-3c55df646cef {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.632126] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1289.632126] env[62914]: value = "task-1353068" [ 1289.632126] env[62914]: _type = "Task" [ 1289.632126] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.640052] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353068, 'name': Rename_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.141839] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353068, 'name': Rename_Task, 'duration_secs': 0.133868} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.142142] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1290.142404] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b283dc7-09af-4a1d-b613-fc444752862b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.148309] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1290.148309] env[62914]: value = "task-1353069" [ 1290.148309] env[62914]: _type = "Task" [ 1290.148309] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.155407] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353069, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.658332] env[62914]: DEBUG oslo_vmware.api [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353069, 'name': PowerOnVM_Task, 'duration_secs': 0.437413} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.658699] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1290.658699] env[62914]: INFO nova.compute.manager [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Took 8.62 seconds to spawn the instance on the hypervisor. [ 1290.658774] env[62914]: DEBUG nova.compute.manager [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1290.659534] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6696b2c9-d753-4868-ab8d-ee26be087af7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.177209] env[62914]: INFO nova.compute.manager [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Took 13.28 seconds to build instance. [ 1291.679615] env[62914]: DEBUG oslo_concurrency.lockutils [None req-b36d99a6-bb47-45fc-adf1-a49affb33ea7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.794s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1291.893215] env[62914]: DEBUG nova.compute.manager [req-e9cf54c7-1adc-4b0e-af21-eee9c99799fb req-497dfe67-5c84-439a-9320-3823213cb93e service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Received event network-changed-4542e979-2f0f-41ec-ab26-d6dbf11ec534 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1291.893425] env[62914]: DEBUG nova.compute.manager [req-e9cf54c7-1adc-4b0e-af21-eee9c99799fb req-497dfe67-5c84-439a-9320-3823213cb93e service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Refreshing instance network info cache due to event network-changed-4542e979-2f0f-41ec-ab26-d6dbf11ec534. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1291.893641] env[62914]: DEBUG oslo_concurrency.lockutils [req-e9cf54c7-1adc-4b0e-af21-eee9c99799fb req-497dfe67-5c84-439a-9320-3823213cb93e service nova] Acquiring lock "refresh_cache-4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1291.893806] env[62914]: DEBUG oslo_concurrency.lockutils [req-e9cf54c7-1adc-4b0e-af21-eee9c99799fb req-497dfe67-5c84-439a-9320-3823213cb93e service nova] Acquired lock "refresh_cache-4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1291.893989] env[62914]: DEBUG nova.network.neutron [req-e9cf54c7-1adc-4b0e-af21-eee9c99799fb req-497dfe67-5c84-439a-9320-3823213cb93e service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Refreshing network info cache for port 4542e979-2f0f-41ec-ab26-d6dbf11ec534 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1292.581489] env[62914]: DEBUG nova.network.neutron [req-e9cf54c7-1adc-4b0e-af21-eee9c99799fb req-497dfe67-5c84-439a-9320-3823213cb93e service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Updated VIF entry in instance network info cache for port 4542e979-2f0f-41ec-ab26-d6dbf11ec534. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1292.581854] env[62914]: DEBUG nova.network.neutron [req-e9cf54c7-1adc-4b0e-af21-eee9c99799fb req-497dfe67-5c84-439a-9320-3823213cb93e service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Updating instance_info_cache with network_info: [{"id": "4542e979-2f0f-41ec-ab26-d6dbf11ec534", "address": "fa:16:3e:c8:6d:06", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4542e979-2f", "ovs_interfaceid": "4542e979-2f0f-41ec-ab26-d6dbf11ec534", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1293.085227] env[62914]: DEBUG oslo_concurrency.lockutils [req-e9cf54c7-1adc-4b0e-af21-eee9c99799fb req-497dfe67-5c84-439a-9320-3823213cb93e service nova] Releasing lock "refresh_cache-4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1311.088806] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1311.089113] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1311.591857] env[62914]: DEBUG nova.compute.manager [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Starting instance... {{(pid=62914) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1312.113774] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1312.114046] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1312.115602] env[62914]: INFO nova.compute.claims [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1313.169734] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df72fae-a4ec-4b5f-89f6-266f6e495f2e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.177676] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b9c9d7a-8707-447a-aaff-b14567ab11ce {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.207756] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b415cf-2184-465b-8207-cb59c01c6ce1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.215223] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ba2dc9-4e30-4c68-a53b-2e289033bffd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.228385] env[62914]: DEBUG nova.compute.provider_tree [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1313.731311] env[62914]: DEBUG nova.scheduler.client.report [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1314.237065] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.122s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1314.237065] env[62914]: DEBUG nova.compute.manager [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Start building networks asynchronously for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1314.742122] env[62914]: DEBUG nova.compute.utils [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Using /dev/sd instead of None {{(pid=62914) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1314.743609] env[62914]: DEBUG nova.compute.manager [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Allocating IP information in the background. {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1314.743806] env[62914]: DEBUG nova.network.neutron [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] allocate_for_instance() {{(pid=62914) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1314.792419] env[62914]: DEBUG nova.policy [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7ab20b70ceb4f5bb700de50494399cc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbb45359291e4d699c21832d0265c1dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62914) authorize /opt/stack/nova/nova/policy.py:201}} [ 1315.041320] env[62914]: DEBUG nova.network.neutron [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Successfully created port: 9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4 {{(pid=62914) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1315.247739] env[62914]: DEBUG nova.compute.manager [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Start building block device mappings for instance. {{(pid=62914) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1315.754065] env[62914]: INFO nova.virt.block_device [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Booting with volume c086247a-eaac-4566-834c-ccd342c62f97 at /dev/sda [ 1315.791537] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a5c40441-b2c9-484c-897b-cfce88dc3d30 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.801209] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4add2c87-beb5-4bac-b8a3-72800a0687be {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.826497] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7a1aa0b1-55ef-41b3-802c-e9aa90f962a4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.835949] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39bd1d8f-721d-4f71-8982-aadefc96d44c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.861394] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109f1f75-51fc-428e-b692-d707893ac1df {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.868099] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-935d82b9-4ae9-44fb-9127-b1d7fd8b7e51 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.882228] env[62914]: DEBUG nova.virt.block_device [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating existing volume attachment record: fe599524-74e2-4532-8628-dce5fdbdaf6b {{(pid=62914) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1316.411375] env[62914]: DEBUG nova.compute.manager [req-555f9aab-f1f1-455c-8818-806abec1d2e5 req-13e25494-acc1-43bd-a96e-bb12e2bf2ad4 service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Received event network-vif-plugged-9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1316.411633] env[62914]: DEBUG oslo_concurrency.lockutils [req-555f9aab-f1f1-455c-8818-806abec1d2e5 req-13e25494-acc1-43bd-a96e-bb12e2bf2ad4 service nova] Acquiring lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1316.411812] env[62914]: DEBUG oslo_concurrency.lockutils [req-555f9aab-f1f1-455c-8818-806abec1d2e5 req-13e25494-acc1-43bd-a96e-bb12e2bf2ad4 service nova] Lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1316.411966] env[62914]: DEBUG oslo_concurrency.lockutils [req-555f9aab-f1f1-455c-8818-806abec1d2e5 req-13e25494-acc1-43bd-a96e-bb12e2bf2ad4 service nova] Lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1316.412160] env[62914]: DEBUG nova.compute.manager [req-555f9aab-f1f1-455c-8818-806abec1d2e5 req-13e25494-acc1-43bd-a96e-bb12e2bf2ad4 service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] No waiting events found dispatching network-vif-plugged-9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4 {{(pid=62914) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1316.412329] env[62914]: WARNING nova.compute.manager [req-555f9aab-f1f1-455c-8818-806abec1d2e5 req-13e25494-acc1-43bd-a96e-bb12e2bf2ad4 service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Received unexpected event network-vif-plugged-9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4 for instance with vm_state building and task_state block_device_mapping. [ 1316.494094] env[62914]: DEBUG nova.network.neutron [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Successfully updated port: 9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4 {{(pid=62914) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1316.997361] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1316.997361] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1316.997361] env[62914]: DEBUG nova.network.neutron [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1317.525895] env[62914]: DEBUG nova.network.neutron [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Instance cache missing network info. {{(pid=62914) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1317.643294] env[62914]: DEBUG nova.network.neutron [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance_info_cache with network_info: [{"id": "9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4", "address": "fa:16:3e:52:8f:84", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b0d7ce2-21", "ovs_interfaceid": "9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1317.962543] env[62914]: DEBUG nova.compute.manager [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Start spawning the instance on the hypervisor. {{(pid=62914) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1317.963104] env[62914]: DEBUG nova.virt.hardware [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1317.963326] env[62914]: DEBUG nova.virt.hardware [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1317.963485] env[62914]: DEBUG nova.virt.hardware [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1317.963671] env[62914]: DEBUG nova.virt.hardware [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1317.963819] env[62914]: DEBUG nova.virt.hardware [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1317.963968] env[62914]: DEBUG nova.virt.hardware [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1317.964192] env[62914]: DEBUG nova.virt.hardware [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1317.964351] env[62914]: DEBUG nova.virt.hardware [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1317.964517] env[62914]: DEBUG nova.virt.hardware [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1317.964736] env[62914]: DEBUG nova.virt.hardware [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1317.964914] env[62914]: DEBUG nova.virt.hardware [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1317.965799] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9372095b-7397-4e76-8a89-27abab4f9cbe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.974253] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba08272-4fd6-426d-a784-6bdfd110865a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.146023] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1318.146376] env[62914]: DEBUG nova.compute.manager [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Instance network_info: |[{"id": "9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4", "address": "fa:16:3e:52:8f:84", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b0d7ce2-21", "ovs_interfaceid": "9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62914) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1318.146829] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:8f:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '504a3211-8288-4b87-a52a-590466d80a49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4', 'vif_model': 'vmxnet3'}] {{(pid=62914) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1318.156502] env[62914]: DEBUG oslo.service.loopingcall [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1318.156727] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Creating VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1318.156989] env[62914]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f68396b9-fcb7-4f4b-aebc-3a403fd85366 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.179107] env[62914]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1318.179107] env[62914]: value = "task-1353078" [ 1318.179107] env[62914]: _type = "Task" [ 1318.179107] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.186853] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1353078, 'name': CreateVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1318.436079] env[62914]: DEBUG nova.compute.manager [req-a9dcceb8-95fa-4411-8f16-9677d35bc280 req-ea12c4ce-22db-4162-9ac2-ea395573b7a7 service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Received event network-changed-9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1318.436236] env[62914]: DEBUG nova.compute.manager [req-a9dcceb8-95fa-4411-8f16-9677d35bc280 req-ea12c4ce-22db-4162-9ac2-ea395573b7a7 service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Refreshing instance network info cache due to event network-changed-9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1318.436460] env[62914]: DEBUG oslo_concurrency.lockutils [req-a9dcceb8-95fa-4411-8f16-9677d35bc280 req-ea12c4ce-22db-4162-9ac2-ea395573b7a7 service nova] Acquiring lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1318.436606] env[62914]: DEBUG oslo_concurrency.lockutils [req-a9dcceb8-95fa-4411-8f16-9677d35bc280 req-ea12c4ce-22db-4162-9ac2-ea395573b7a7 service nova] Acquired lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1318.436772] env[62914]: DEBUG nova.network.neutron [req-a9dcceb8-95fa-4411-8f16-9677d35bc280 req-ea12c4ce-22db-4162-9ac2-ea395573b7a7 service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Refreshing network info cache for port 9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1318.690378] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1353078, 'name': CreateVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.119464] env[62914]: DEBUG nova.network.neutron [req-a9dcceb8-95fa-4411-8f16-9677d35bc280 req-ea12c4ce-22db-4162-9ac2-ea395573b7a7 service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updated VIF entry in instance network info cache for port 9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1319.119857] env[62914]: DEBUG nova.network.neutron [req-a9dcceb8-95fa-4411-8f16-9677d35bc280 req-ea12c4ce-22db-4162-9ac2-ea395573b7a7 service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance_info_cache with network_info: [{"id": "9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4", "address": "fa:16:3e:52:8f:84", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b0d7ce2-21", "ovs_interfaceid": "9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1319.191245] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1353078, 'name': CreateVM_Task} progress is 99%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.624731] env[62914]: DEBUG oslo_concurrency.lockutils [req-a9dcceb8-95fa-4411-8f16-9677d35bc280 req-ea12c4ce-22db-4162-9ac2-ea395573b7a7 service nova] Releasing lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1319.691421] env[62914]: DEBUG oslo_vmware.api [-] Task: {'id': task-1353078, 'name': CreateVM_Task, 'duration_secs': 1.324568} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1319.691756] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Created VM on the ESX host {{(pid=62914) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1319.692266] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': True, 'attachment_id': 'fe599524-74e2-4532-8628-dce5fdbdaf6b', 'disk_bus': None, 'guest_format': None, 'mount_device': '/dev/sda', 'device_type': None, 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288283', 'volume_id': 'c086247a-eaac-4566-834c-ccd342c62f97', 'name': 'volume-c086247a-eaac-4566-834c-ccd342c62f97', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '52e683d3-c2a5-493d-84ce-1ccdc0ce470b', 'attached_at': '', 'detached_at': '', 'volume_id': 'c086247a-eaac-4566-834c-ccd342c62f97', 'serial': 'c086247a-eaac-4566-834c-ccd342c62f97'}, 'volume_type': None}], 'swap': None} {{(pid=62914) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1319.692479] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Root volume attach. Driver type: vmdk {{(pid=62914) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1319.693240] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17dc1456-d9cd-48a2-8ce0-bbc89d46f1b0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.700535] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4a8334-2d1a-495b-b949-737410d854bd {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.706337] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493bc503-fd9b-45d6-9ee7-6311575d4fa0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.711796] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-82a15afe-4f58-4899-bf08-ce966da6d8e6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.720307] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1319.720307] env[62914]: value = "task-1353079" [ 1319.720307] env[62914]: _type = "Task" [ 1319.720307] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.727310] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353079, 'name': RelocateVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.230727] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353079, 'name': RelocateVM_Task, 'duration_secs': 0.339372} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.231045] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Volume attach. Driver type: vmdk {{(pid=62914) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1320.231265] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288283', 'volume_id': 'c086247a-eaac-4566-834c-ccd342c62f97', 'name': 'volume-c086247a-eaac-4566-834c-ccd342c62f97', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '52e683d3-c2a5-493d-84ce-1ccdc0ce470b', 'attached_at': '', 'detached_at': '', 'volume_id': 'c086247a-eaac-4566-834c-ccd342c62f97', 'serial': 'c086247a-eaac-4566-834c-ccd342c62f97'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1320.232027] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb89a1aa-3938-4ad6-932b-398764b58be7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.247507] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32efe7c9-a595-4369-87eb-32acf1179fbc {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.268742] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] volume-c086247a-eaac-4566-834c-ccd342c62f97/volume-c086247a-eaac-4566-834c-ccd342c62f97.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1320.269012] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b269ea7b-c91e-451b-88e7-0842b2d692b9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.288799] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1320.288799] env[62914]: value = "task-1353080" [ 1320.288799] env[62914]: _type = "Task" [ 1320.288799] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.296953] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353080, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.799329] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353080, 'name': ReconfigVM_Task, 'duration_secs': 0.254497} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.799732] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Reconfigured VM instance instance-0000006c to attach disk [datastore2] volume-c086247a-eaac-4566-834c-ccd342c62f97/volume-c086247a-eaac-4566-834c-ccd342c62f97.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1320.804301] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd5afebf-033f-412e-9460-ced94351f832 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.819636] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1320.819636] env[62914]: value = "task-1353081" [ 1320.819636] env[62914]: _type = "Task" [ 1320.819636] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.828571] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353081, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.330061] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353081, 'name': ReconfigVM_Task, 'duration_secs': 0.123664} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.330374] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288283', 'volume_id': 'c086247a-eaac-4566-834c-ccd342c62f97', 'name': 'volume-c086247a-eaac-4566-834c-ccd342c62f97', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '52e683d3-c2a5-493d-84ce-1ccdc0ce470b', 'attached_at': '', 'detached_at': '', 'volume_id': 'c086247a-eaac-4566-834c-ccd342c62f97', 'serial': 'c086247a-eaac-4566-834c-ccd342c62f97'} {{(pid=62914) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1321.330911] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6f106329-08ff-4f63-8b7f-b95912fc58a6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.338547] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1321.338547] env[62914]: value = "task-1353082" [ 1321.338547] env[62914]: _type = "Task" [ 1321.338547] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.346454] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353082, 'name': Rename_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.849205] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353082, 'name': Rename_Task, 'duration_secs': 0.133619} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.849575] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1321.849725] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ea336ad-6f8e-46b2-a6fe-4a71be1bd4a9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.856525] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1321.856525] env[62914]: value = "task-1353083" [ 1321.856525] env[62914]: _type = "Task" [ 1321.856525] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.863658] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353083, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.366579] env[62914]: DEBUG oslo_vmware.api [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353083, 'name': PowerOnVM_Task, 'duration_secs': 0.420971} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1322.366852] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1322.367074] env[62914]: INFO nova.compute.manager [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Took 4.40 seconds to spawn the instance on the hypervisor. [ 1322.367264] env[62914]: DEBUG nova.compute.manager [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Checking state {{(pid=62914) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1322.367998] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5407d548-4410-4dab-b2d9-e68305b80871 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.884961] env[62914]: INFO nova.compute.manager [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Took 10.79 seconds to build instance. [ 1323.178034] env[62914]: DEBUG nova.compute.manager [req-faee62bf-8bca-49cc-8f03-6aa0034aace8 req-ff90c474-375a-4d70-b277-89a322c408eb service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Received event network-changed-e4f0d675-9d9c-4f76-bc5b-10ce73f445bf {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1323.178249] env[62914]: DEBUG nova.compute.manager [req-faee62bf-8bca-49cc-8f03-6aa0034aace8 req-ff90c474-375a-4d70-b277-89a322c408eb service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Refreshing instance network info cache due to event network-changed-e4f0d675-9d9c-4f76-bc5b-10ce73f445bf. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1323.178473] env[62914]: DEBUG oslo_concurrency.lockutils [req-faee62bf-8bca-49cc-8f03-6aa0034aace8 req-ff90c474-375a-4d70-b277-89a322c408eb service nova] Acquiring lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1323.178620] env[62914]: DEBUG oslo_concurrency.lockutils [req-faee62bf-8bca-49cc-8f03-6aa0034aace8 req-ff90c474-375a-4d70-b277-89a322c408eb service nova] Acquired lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1323.178783] env[62914]: DEBUG nova.network.neutron [req-faee62bf-8bca-49cc-8f03-6aa0034aace8 req-ff90c474-375a-4d70-b277-89a322c408eb service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Refreshing network info cache for port e4f0d675-9d9c-4f76-bc5b-10ce73f445bf {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1323.387396] env[62914]: DEBUG oslo_concurrency.lockutils [None req-3cbde25b-0a97-4617-8be7-eb89e3994f9f tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.298s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1323.889410] env[62914]: DEBUG nova.network.neutron [req-faee62bf-8bca-49cc-8f03-6aa0034aace8 req-ff90c474-375a-4d70-b277-89a322c408eb service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updated VIF entry in instance network info cache for port e4f0d675-9d9c-4f76-bc5b-10ce73f445bf. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1323.889775] env[62914]: DEBUG nova.network.neutron [req-faee62bf-8bca-49cc-8f03-6aa0034aace8 req-ff90c474-375a-4d70-b277-89a322c408eb service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updating instance_info_cache with network_info: [{"id": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "address": "fa:16:3e:fb:87:48", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4f0d675-9d", "ovs_interfaceid": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1324.392763] env[62914]: DEBUG oslo_concurrency.lockutils [req-faee62bf-8bca-49cc-8f03-6aa0034aace8 req-ff90c474-375a-4d70-b277-89a322c408eb service nova] Releasing lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1324.971482] env[62914]: DEBUG nova.compute.manager [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Stashing vm_state: active {{(pid=62914) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1325.204540] env[62914]: DEBUG nova.compute.manager [req-e565780c-a356-4fbc-a6e4-d75515021e12 req-1d23978f-a625-4f20-98aa-de7a51909026 service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Received event network-changed-9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1325.204649] env[62914]: DEBUG nova.compute.manager [req-e565780c-a356-4fbc-a6e4-d75515021e12 req-1d23978f-a625-4f20-98aa-de7a51909026 service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Refreshing instance network info cache due to event network-changed-9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4. {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1325.204946] env[62914]: DEBUG oslo_concurrency.lockutils [req-e565780c-a356-4fbc-a6e4-d75515021e12 req-1d23978f-a625-4f20-98aa-de7a51909026 service nova] Acquiring lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1325.205142] env[62914]: DEBUG oslo_concurrency.lockutils [req-e565780c-a356-4fbc-a6e4-d75515021e12 req-1d23978f-a625-4f20-98aa-de7a51909026 service nova] Acquired lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1325.205328] env[62914]: DEBUG nova.network.neutron [req-e565780c-a356-4fbc-a6e4-d75515021e12 req-1d23978f-a625-4f20-98aa-de7a51909026 service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Refreshing network info cache for port 9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4 {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1325.492521] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1325.492832] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1325.919572] env[62914]: DEBUG nova.network.neutron [req-e565780c-a356-4fbc-a6e4-d75515021e12 req-1d23978f-a625-4f20-98aa-de7a51909026 service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updated VIF entry in instance network info cache for port 9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4. {{(pid=62914) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1325.919943] env[62914]: DEBUG nova.network.neutron [req-e565780c-a356-4fbc-a6e4-d75515021e12 req-1d23978f-a625-4f20-98aa-de7a51909026 service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance_info_cache with network_info: [{"id": "9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4", "address": "fa:16:3e:52:8f:84", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b0d7ce2-21", "ovs_interfaceid": "9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1325.998159] env[62914]: INFO nova.compute.claims [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1326.423096] env[62914]: DEBUG oslo_concurrency.lockutils [req-e565780c-a356-4fbc-a6e4-d75515021e12 req-1d23978f-a625-4f20-98aa-de7a51909026 service nova] Releasing lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1326.504206] env[62914]: INFO nova.compute.resource_tracker [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating resource usage from migration 4b016131-99b0-4051-9089-43d65097adb4 [ 1326.560176] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08afe17d-cc8f-48db-9300-d23435341525 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.567695] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adce1476-8b26-4ecc-931f-722957884af1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.596269] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baaf1fb9-4580-460d-ad4f-25f4b9a55e74 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.603083] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46bab0a1-e895-49f9-9228-770e330d7a76 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.616756] env[62914]: DEBUG nova.compute.provider_tree [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1327.120080] env[62914]: DEBUG nova.scheduler.client.report [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1327.624661] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.132s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1327.624924] env[62914]: INFO nova.compute.manager [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Migrating [ 1328.139918] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1328.140276] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1328.140357] env[62914]: DEBUG nova.network.neutron [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1328.343058] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1328.343225] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1328.832137] env[62914]: DEBUG nova.network.neutron [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance_info_cache with network_info: [{"id": "9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4", "address": "fa:16:3e:52:8f:84", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b0d7ce2-21", "ovs_interfaceid": "9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1328.847359] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1328.847548] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Starting heal instance info cache {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1329.334569] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1329.349736] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Didn't find any instances for network info cache update. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1329.350160] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.350362] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.350482] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.350628] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.350767] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.350908] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.351246] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62914) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1329.351246] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager.update_available_resource {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.853655] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1329.853938] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1329.854056] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1329.854217] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62914) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1329.855195] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07aab41-9176-4cde-9e01-b64955d0cd56 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.863809] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9922eb0e-dfed-4d8e-ae99-455d745134c8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.879395] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af0538f9-60bb-43b1-9f3b-7b4ab1865fe0 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.886358] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba49e56-0710-4367-b38e-81b157c03053 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.915709] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181106MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62914) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1329.915877] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1329.916073] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.849573] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8906e09d-132c-4dd2-8421-976e752d2b6c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.867555] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance '52e683d3-c2a5-493d-84ce-1ccdc0ce470b' progress to 0 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1330.923416] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Applying migration context for instance 52e683d3-c2a5-493d-84ce-1ccdc0ce470b as it has an incoming, in-progress migration 4b016131-99b0-4051-9089-43d65097adb4. Migration status is migrating {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1330.924115] env[62914]: INFO nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating resource usage from migration 4b016131-99b0-4051-9089-43d65097adb4 [ 1330.938543] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance f1c5c877-9c29-4331-84f8-23b3c6b53725 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1330.938681] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1330.938802] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Migration 4b016131-99b0-4051-9089-43d65097adb4 is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1330.938918] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 52e683d3-c2a5-493d-84ce-1ccdc0ce470b actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1330.939112] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1330.939235] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1330.989055] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5aa3c24-3fb6-4c6d-a784-1a6fe11115ac {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.996353] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ba32cf-0da3-46e5-a661-e57a11459a8a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.026252] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4fab98e-096d-4ff9-8359-3506154cfb2b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.033012] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3954e962-0deb-4105-bc49-6d882867adfa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.046037] env[62914]: DEBUG nova.compute.provider_tree [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1331.373344] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1331.373645] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af828cc3-1316-4756-8cba-c910e560c6fa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.381165] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1331.381165] env[62914]: value = "task-1353084" [ 1331.381165] env[62914]: _type = "Task" [ 1331.381165] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.388655] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353084, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.548944] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1331.890557] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353084, 'name': PowerOffVM_Task, 'duration_secs': 0.217603} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.890948] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1331.891046] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance '52e683d3-c2a5-493d-84ce-1ccdc0ce470b' progress to 17 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1332.053532] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1332.053724] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.138s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1332.397648] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1332.397889] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1332.398063] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1332.398253] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1332.398402] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1332.398557] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1332.398768] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1332.398935] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1332.399122] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1332.399290] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1332.399460] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1332.404516] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f43511f2-e24a-46f9-a08e-4b5246ebcb8e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.420774] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1332.420774] env[62914]: value = "task-1353085" [ 1332.420774] env[62914]: _type = "Task" [ 1332.420774] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.428667] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353085, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.931017] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353085, 'name': ReconfigVM_Task, 'duration_secs': 0.143093} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1332.931396] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance '52e683d3-c2a5-493d-84ce-1ccdc0ce470b' progress to 33 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1333.437410] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T12:07:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1333.437701] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1333.437860] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image limits 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1333.438058] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Flavor pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1333.438214] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Image pref 0:0:0 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1333.438364] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62914) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1333.438573] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1333.438733] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1333.438972] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Got 1 possible topologies {{(pid=62914) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1333.439189] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1333.439370] env[62914]: DEBUG nova.virt.hardware [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62914) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1333.444631] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1333.444964] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11fed0b7-9e21-4938-a1f5-69e7b693382b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.463754] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1333.463754] env[62914]: value = "task-1353086" [ 1333.463754] env[62914]: _type = "Task" [ 1333.463754] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.472650] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353086, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.976284] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353086, 'name': ReconfigVM_Task, 'duration_secs': 0.150616} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1333.976580] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1333.977350] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de5cb7a-1992-44a1-8bd6-e1a0a143d7eb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.998238] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] volume-c086247a-eaac-4566-834c-ccd342c62f97/volume-c086247a-eaac-4566-834c-ccd342c62f97.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1333.998760] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cfe2b283-bebf-409a-ae62-65020b368b7b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.016321] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1334.016321] env[62914]: value = "task-1353087" [ 1334.016321] env[62914]: _type = "Task" [ 1334.016321] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.023854] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353087, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.526054] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353087, 'name': ReconfigVM_Task, 'duration_secs': 0.250058} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.526054] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Reconfigured VM instance instance-0000006c to attach disk [datastore2] volume-c086247a-eaac-4566-834c-ccd342c62f97/volume-c086247a-eaac-4566-834c-ccd342c62f97.vmdk or device None with type thin {{(pid=62914) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1334.526246] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance '52e683d3-c2a5-493d-84ce-1ccdc0ce470b' progress to 50 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1335.033280] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52bfee48-2e43-49d0-9822-fbce9c47213c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.051623] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0a3f29-aa60-4028-9a87-a8e2310ce9c6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.068066] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance '52e683d3-c2a5-493d-84ce-1ccdc0ce470b' progress to 67 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1336.714091] env[62914]: DEBUG nova.network.neutron [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Port 9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4 binding to destination host cpu-1 is already ACTIVE {{(pid=62914) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1337.735380] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1337.735812] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1337.735951] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1338.770078] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1338.770457] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1338.770457] env[62914]: DEBUG nova.network.neutron [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1339.470012] env[62914]: DEBUG nova.network.neutron [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance_info_cache with network_info: [{"id": "9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4", "address": "fa:16:3e:52:8f:84", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b0d7ce2-21", "ovs_interfaceid": "9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1339.973272] env[62914]: DEBUG oslo_concurrency.lockutils [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1340.482549] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e169e338-4243-4086-8d92-e68250018480 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.489608] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68511640-7404-4656-8d9e-054366c9e654 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.588870] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a01caf8-4ae7-4ce3-84f4-f39c79a7a038 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.608530] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b68eb4c-349d-43fe-8712-307a25a0b01c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.615308] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance '52e683d3-c2a5-493d-84ce-1ccdc0ce470b' progress to 83 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1342.121940] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Powering on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1342.122295] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7c07d38-c176-4b9a-9216-48da7707b537 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.129825] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1342.129825] env[62914]: value = "task-1353088" [ 1342.129825] env[62914]: _type = "Task" [ 1342.129825] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1342.137669] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353088, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.639580] env[62914]: DEBUG oslo_vmware.api [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353088, 'name': PowerOnVM_Task, 'duration_secs': 0.37957} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.640825] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Powered on the VM {{(pid=62914) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1342.640825] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-bf8ae5a2-f5fe-49c8-9e4f-e31fa4cacb8c tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance '52e683d3-c2a5-493d-84ce-1ccdc0ce470b' progress to 100 {{(pid=62914) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1345.367194] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1345.367572] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1345.367702] env[62914]: DEBUG nova.compute.manager [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Going to confirm migration 3 {{(pid=62914) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1345.902380] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1345.902580] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquired lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1345.902760] env[62914]: DEBUG nova.network.neutron [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Building network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1345.902958] env[62914]: DEBUG nova.objects.instance [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lazy-loading 'info_cache' on Instance uuid 52e683d3-c2a5-493d-84ce-1ccdc0ce470b {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1347.101305] env[62914]: DEBUG nova.network.neutron [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance_info_cache with network_info: [{"id": "9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4", "address": "fa:16:3e:52:8f:84", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b0d7ce2-21", "ovs_interfaceid": "9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1347.604609] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Releasing lock "refresh_cache-52e683d3-c2a5-493d-84ce-1ccdc0ce470b" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1347.604899] env[62914]: DEBUG nova.objects.instance [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lazy-loading 'migration_context' on Instance uuid 52e683d3-c2a5-493d-84ce-1ccdc0ce470b {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1348.108601] env[62914]: DEBUG nova.objects.base [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Object Instance<52e683d3-c2a5-493d-84ce-1ccdc0ce470b> lazy-loaded attributes: info_cache,migration_context {{(pid=62914) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1348.109559] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66efa93c-1b7e-4f11-bedc-dbd6226c92ce {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.128936] env[62914]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c91b84e0-f0a2-4950-a2b8-c234e8feafa1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.134092] env[62914]: DEBUG oslo_vmware.api [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1348.134092] env[62914]: value = "session[52bbc43a-1940-d81b-6922-9c89e07626cc]522ab773-b7d6-295a-97d8-392f541903a4" [ 1348.134092] env[62914]: _type = "Task" [ 1348.134092] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1348.141447] env[62914]: DEBUG oslo_vmware.api [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]522ab773-b7d6-295a-97d8-392f541903a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1348.644220] env[62914]: DEBUG oslo_vmware.api [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': session[52bbc43a-1940-d81b-6922-9c89e07626cc]522ab773-b7d6-295a-97d8-392f541903a4, 'name': SearchDatastore_Task, 'duration_secs': 0.008084} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.644510] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1348.644738] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1349.213788] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f098de7-c01f-4533-afe3-98e19c21bf4d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.222084] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10076957-bfd5-43cf-9d79-59d19f96a852 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.251386] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c033f1a-71e5-4b0f-9975-5c343c835c60 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.258949] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-059f4ef5-05a9-44ad-bbcf-fc2db0c28ed7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.273723] env[62914]: DEBUG nova.compute.provider_tree [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1349.777225] env[62914]: DEBUG nova.scheduler.client.report [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1350.788732] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.143s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1351.344784] env[62914]: INFO nova.scheduler.client.report [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleted allocation for migration 4b016131-99b0-4051-9089-43d65097adb4 [ 1351.699536] env[62914]: INFO nova.compute.manager [None req-74d42e8f-b126-40d4-9b55-1db7f90f5c8a tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Get console output [ 1351.699924] env[62914]: WARNING nova.virt.vmwareapi.driver [None req-74d42e8f-b126-40d4-9b55-1db7f90f5c8a tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] The console log is missing. Check your VSPC configuration [ 1351.850305] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e7e077c9-7f8b-45e4-acd8-ba38a673cee7 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.483s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1371.902741] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1371.903170] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Cleaning up deleted instances {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1372.407731] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] There are 6 instances to clean {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1372.407984] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 0e461e81-ec47-40b6-9c09-87037f628750] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1372.911533] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 497c4c93-801e-41a6-ba8d-1c2f51710546] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1373.414962] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 1005f4b0-e07a-4f32-9b3d-a2868a307bb9] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1373.917972] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 58c2cc4a-153d-437b-8822-d3e389bd1a7f] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1374.421026] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 1028b121-ea93-4d9a-b2fd-0c2483c5c618] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1374.924408] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 062cd4d7-3cf1-4f31-9631-e9d1e194ad53] Instance has had 0 of 5 cleanup attempts {{(pid=62914) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1378.595708] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1378.596124] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1378.596231] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1378.596421] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1378.596592] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1378.599776] env[62914]: INFO nova.compute.manager [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Terminating instance [ 1378.601722] env[62914]: DEBUG nova.compute.manager [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1378.602241] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1378.602241] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e1ebca1-910f-4f77-821a-571219ad9ab7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.609085] env[62914]: DEBUG oslo_vmware.api [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1378.609085] env[62914]: value = "task-1353089" [ 1378.609085] env[62914]: _type = "Task" [ 1378.609085] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1378.618442] env[62914]: DEBUG oslo_vmware.api [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353089, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1379.118717] env[62914]: DEBUG oslo_vmware.api [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353089, 'name': PowerOffVM_Task, 'duration_secs': 0.155651} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1379.118989] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1379.119206] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Volume detach. Driver type: vmdk {{(pid=62914) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1379.119405] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288283', 'volume_id': 'c086247a-eaac-4566-834c-ccd342c62f97', 'name': 'volume-c086247a-eaac-4566-834c-ccd342c62f97', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '52e683d3-c2a5-493d-84ce-1ccdc0ce470b', 'attached_at': '2024-10-24T12:22:38.000000', 'detached_at': '', 'volume_id': 'c086247a-eaac-4566-834c-ccd342c62f97', 'serial': 'c086247a-eaac-4566-834c-ccd342c62f97'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1379.120163] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-026162ce-54ed-4e55-ad5f-77b247eb9e0d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.138128] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b23c08-bff4-4d09-b51e-7da7c9ae4f69 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.144156] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ca9166-4667-4367-9996-79e741aa929c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.160671] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a61767-d2e6-4701-9501-19f25af1cab2 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.175446] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] The volume has not been displaced from its original location: [datastore2] volume-c086247a-eaac-4566-834c-ccd342c62f97/volume-c086247a-eaac-4566-834c-ccd342c62f97.vmdk. No consolidation needed. {{(pid=62914) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1379.180647] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1379.180900] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03cad75c-9988-47de-8f8a-56bc30593acb {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.198136] env[62914]: DEBUG oslo_vmware.api [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1379.198136] env[62914]: value = "task-1353090" [ 1379.198136] env[62914]: _type = "Task" [ 1379.198136] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1379.205185] env[62914]: DEBUG oslo_vmware.api [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353090, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1379.708146] env[62914]: DEBUG oslo_vmware.api [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353090, 'name': ReconfigVM_Task, 'duration_secs': 0.151552} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1379.708537] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=62914) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1379.712854] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8241bd8-b19d-457f-a9a9-bab2e3c8c344 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.728201] env[62914]: DEBUG oslo_vmware.api [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1379.728201] env[62914]: value = "task-1353091" [ 1379.728201] env[62914]: _type = "Task" [ 1379.728201] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1379.737420] env[62914]: DEBUG oslo_vmware.api [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353091, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1380.238014] env[62914]: DEBUG oslo_vmware.api [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353091, 'name': ReconfigVM_Task, 'duration_secs': 0.147261} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1380.238343] env[62914]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288283', 'volume_id': 'c086247a-eaac-4566-834c-ccd342c62f97', 'name': 'volume-c086247a-eaac-4566-834c-ccd342c62f97', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '52e683d3-c2a5-493d-84ce-1ccdc0ce470b', 'attached_at': '2024-10-24T12:22:38.000000', 'detached_at': '', 'volume_id': 'c086247a-eaac-4566-834c-ccd342c62f97', 'serial': 'c086247a-eaac-4566-834c-ccd342c62f97'} {{(pid=62914) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1380.238615] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1380.239371] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5faf6d2-ce86-428f-b9b6-5b66c72ed8b3 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.245395] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1380.245644] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56a62603-31bf-48b1-b26f-bd5180f1b342 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.303815] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1380.304054] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1380.304247] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleting the datastore file [datastore2] 52e683d3-c2a5-493d-84ce-1ccdc0ce470b {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1380.304504] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d4ce872e-38f3-4288-a9ab-6c569b55f9a4 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.310813] env[62914]: DEBUG oslo_vmware.api [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1380.310813] env[62914]: value = "task-1353093" [ 1380.310813] env[62914]: _type = "Task" [ 1380.310813] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1380.318084] env[62914]: DEBUG oslo_vmware.api [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353093, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1380.426813] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1380.427071] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1380.427220] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Starting heal instance info cache {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1380.427337] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Rebuilding the list of instances to heal {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1380.821181] env[62914]: DEBUG oslo_vmware.api [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353093, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.084309} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1380.821604] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1380.821604] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1380.821794] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1380.821972] env[62914]: INFO nova.compute.manager [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Took 2.22 seconds to destroy the instance on the hypervisor. [ 1380.822229] env[62914]: DEBUG oslo.service.loopingcall [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1380.822426] env[62914]: DEBUG nova.compute.manager [-] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1380.822524] env[62914]: DEBUG nova.network.neutron [-] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1380.931380] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Skipping network cache update for instance because it is being deleted. {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1380.957569] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1380.957743] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquired lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1380.957927] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Forcefully refreshing network info cache for instance {{(pid=62914) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1380.958097] env[62914]: DEBUG nova.objects.instance [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lazy-loading 'info_cache' on Instance uuid f1c5c877-9c29-4331-84f8-23b3c6b53725 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1381.226400] env[62914]: DEBUG nova.compute.manager [req-b8042d3d-d074-43b5-82b7-d64263de7a3f req-0b2dd225-cdae-4e72-b7ac-faee90cf390b service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Received event network-vif-deleted-9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1381.226400] env[62914]: INFO nova.compute.manager [req-b8042d3d-d074-43b5-82b7-d64263de7a3f req-0b2dd225-cdae-4e72-b7ac-faee90cf390b service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Neutron deleted interface 9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4; detaching it from the instance and deleting it from the info cache [ 1381.226400] env[62914]: DEBUG nova.network.neutron [req-b8042d3d-d074-43b5-82b7-d64263de7a3f req-0b2dd225-cdae-4e72-b7ac-faee90cf390b service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1381.705926] env[62914]: DEBUG nova.network.neutron [-] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1381.729156] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c07134f5-ef4f-48b5-a245-d9dc2da82249 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.740708] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce860c48-e08d-41db-bad8-80e083958e08 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.764741] env[62914]: DEBUG nova.compute.manager [req-b8042d3d-d074-43b5-82b7-d64263de7a3f req-0b2dd225-cdae-4e72-b7ac-faee90cf390b service nova] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Detach interface failed, port_id=9b0d7ce2-216f-46d5-a42c-4d0f37dcd2e4, reason: Instance 52e683d3-c2a5-493d-84ce-1ccdc0ce470b could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1382.209195] env[62914]: INFO nova.compute.manager [-] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Took 1.39 seconds to deallocate network for instance. [ 1382.664167] env[62914]: DEBUG nova.network.neutron [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updating instance_info_cache with network_info: [{"id": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "address": "fa:16:3e:fb:87:48", "network": {"id": "a4500e7a-505d-4a85-a82b-707a1b8b19be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1397281149-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbb45359291e4d699c21832d0265c1dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4f0d675-9d", "ovs_interfaceid": "e4f0d675-9d9c-4f76-bc5b-10ce73f445bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1382.752062] env[62914]: INFO nova.compute.manager [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Took 0.54 seconds to detach 1 volumes for instance. [ 1382.754183] env[62914]: DEBUG nova.compute.manager [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 52e683d3-c2a5-493d-84ce-1ccdc0ce470b] Deleting volume: c086247a-eaac-4566-834c-ccd342c62f97 {{(pid=62914) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1383.166814] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Releasing lock "refresh_cache-f1c5c877-9c29-4331-84f8-23b3c6b53725" {{(pid=62914) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1383.167062] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updated the network info_cache for instance {{(pid=62914) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1383.167194] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1383.167350] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1383.167492] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1383.167636] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1383.167772] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1383.167913] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1383.168049] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62914) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1383.168191] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager.update_available_resource {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1383.292711] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1383.293092] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1383.293176] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1383.323765] env[62914]: INFO nova.scheduler.client.report [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleted allocations for instance 52e683d3-c2a5-493d-84ce-1ccdc0ce470b [ 1383.671361] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1383.671581] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1383.671749] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1383.671900] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62914) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1383.673011] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6cefac9-9bb5-4c81-a5eb-e5a39f1affea {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.681656] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc2a3a1-705d-417f-996b-abc01be6d7d5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.695309] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3584106b-f97d-4614-828c-0da7d58f6efe {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.701417] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fd7f55-7e8f-488a-8047-fa90e1c54b6e {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.734112] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181106MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62914) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1383.734320] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1383.734480] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1383.831102] env[62914]: DEBUG oslo_concurrency.lockutils [None req-e2f9d7d8-a001-4839-ae7e-e15d35214dc9 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "52e683d3-c2a5-493d-84ce-1ccdc0ce470b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.235s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1384.161611] env[62914]: DEBUG oslo_concurrency.lockutils [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1384.161879] env[62914]: DEBUG oslo_concurrency.lockutils [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1384.162112] env[62914]: DEBUG oslo_concurrency.lockutils [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "4ee2aef3-a5ca-4049-88d5-0b2b0817bce7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1384.162300] env[62914]: DEBUG oslo_concurrency.lockutils [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "4ee2aef3-a5ca-4049-88d5-0b2b0817bce7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1384.162471] env[62914]: DEBUG oslo_concurrency.lockutils [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "4ee2aef3-a5ca-4049-88d5-0b2b0817bce7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1384.164625] env[62914]: INFO nova.compute.manager [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Terminating instance [ 1384.166414] env[62914]: DEBUG nova.compute.manager [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1384.166620] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1384.167448] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54220e41-3cac-42d6-8fec-423bc45b1a0d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.174950] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1384.175189] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa132af5-619b-47cc-997e-6a7d74e39197 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.181052] env[62914]: DEBUG oslo_vmware.api [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1384.181052] env[62914]: value = "task-1353095" [ 1384.181052] env[62914]: _type = "Task" [ 1384.181052] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1384.188635] env[62914]: DEBUG oslo_vmware.api [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353095, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1384.690442] env[62914]: DEBUG oslo_vmware.api [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353095, 'name': PowerOffVM_Task, 'duration_secs': 0.194977} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1384.690766] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1384.690869] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1384.691136] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d12b8588-1b89-4f83-8dd6-1cd5d0361d9a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.752911] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1384.753155] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Deleting contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1384.753343] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleting the datastore file [datastore1] 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1384.753656] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca598c2b-f80c-42e6-98c5-42727c869d82 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.757017] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance f1c5c877-9c29-4331-84f8-23b3c6b53725 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1384.757170] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Instance 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62914) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1384.757344] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1384.757481] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62914) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1384.760922] env[62914]: DEBUG oslo_vmware.api [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1384.760922] env[62914]: value = "task-1353097" [ 1384.760922] env[62914]: _type = "Task" [ 1384.760922] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1384.769494] env[62914]: DEBUG oslo_vmware.api [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353097, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1384.792274] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c29f60-00a4-4206-b4e0-2ca051e62c78 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.798965] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f950146-3cb5-40dc-8840-de3320de6b3a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.828729] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f19e764-6744-4e86-a3f8-497ea3af245b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.835578] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ab78c7-4c5d-4cd5-99c4-44dff90c7f3d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.848611] env[62914]: DEBUG nova.compute.provider_tree [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1385.272197] env[62914]: DEBUG oslo_vmware.api [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353097, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142045} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1385.272623] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1385.272921] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Deleted contents of the VM from datastore datastore1 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1385.273234] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1385.273514] env[62914]: INFO nova.compute.manager [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1385.273772] env[62914]: DEBUG oslo.service.loopingcall [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1385.273972] env[62914]: DEBUG nova.compute.manager [-] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1385.274082] env[62914]: DEBUG nova.network.neutron [-] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1385.352669] env[62914]: DEBUG nova.scheduler.client.report [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1385.518153] env[62914]: DEBUG nova.compute.manager [req-d2566506-410f-4879-95ed-30aa72a51bc4 req-4fce6b64-d48f-454a-9e55-c5ad68a71502 service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Received event network-vif-deleted-4542e979-2f0f-41ec-ab26-d6dbf11ec534 {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1385.518345] env[62914]: INFO nova.compute.manager [req-d2566506-410f-4879-95ed-30aa72a51bc4 req-4fce6b64-d48f-454a-9e55-c5ad68a71502 service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Neutron deleted interface 4542e979-2f0f-41ec-ab26-d6dbf11ec534; detaching it from the instance and deleting it from the info cache [ 1385.518521] env[62914]: DEBUG nova.network.neutron [req-d2566506-410f-4879-95ed-30aa72a51bc4 req-4fce6b64-d48f-454a-9e55-c5ad68a71502 service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1385.857096] env[62914]: DEBUG nova.compute.resource_tracker [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62914) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1385.857468] env[62914]: DEBUG oslo_concurrency.lockutils [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.123s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1385.902200] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1385.996640] env[62914]: DEBUG nova.network.neutron [-] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1386.021681] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0de4a7b9-2bfb-4f28-a81a-6ecf5ddb2dc9 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.031095] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a80ffedb-515a-44cd-8798-09f2105abb7d {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.052534] env[62914]: DEBUG nova.compute.manager [req-d2566506-410f-4879-95ed-30aa72a51bc4 req-4fce6b64-d48f-454a-9e55-c5ad68a71502 service nova] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Detach interface failed, port_id=4542e979-2f0f-41ec-ab26-d6dbf11ec534, reason: Instance 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1386.499020] env[62914]: INFO nova.compute.manager [-] [instance: 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7] Took 1.22 seconds to deallocate network for instance. [ 1387.005510] env[62914]: DEBUG oslo_concurrency.lockutils [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1387.005816] env[62914]: DEBUG oslo_concurrency.lockutils [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1387.006062] env[62914]: DEBUG nova.objects.instance [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lazy-loading 'resources' on Instance uuid 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1387.404634] env[62914]: DEBUG oslo_service.periodic_task [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62914) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1387.404823] env[62914]: DEBUG nova.compute.manager [None req-154357d9-d844-40dd-97be-7138ae3741fc None None] Cleaning up deleted instances with incomplete migration {{(pid=62914) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1387.614260] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052bc5f3-3985-4b49-95fb-ed4f11923eaa {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.621489] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505b1fec-6007-4ce8-9546-04778e306c83 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.650565] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d13b727-7f60-44e1-bd39-bee83a269bf1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.657202] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd98a591-1739-4076-abd5-5d59a74a607b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.669606] env[62914]: DEBUG nova.compute.provider_tree [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1388.172728] env[62914]: DEBUG nova.scheduler.client.report [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1388.677768] env[62914]: DEBUG oslo_concurrency.lockutils [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.672s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1388.695069] env[62914]: INFO nova.scheduler.client.report [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleted allocations for instance 4ee2aef3-a5ca-4049-88d5-0b2b0817bce7 [ 1389.202446] env[62914]: DEBUG oslo_concurrency.lockutils [None req-84d3c739-4644-41b6-9595-9b04cfac3984 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "4ee2aef3-a5ca-4049-88d5-0b2b0817bce7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.040s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1392.781798] env[62914]: DEBUG oslo_concurrency.lockutils [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "f1c5c877-9c29-4331-84f8-23b3c6b53725" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1392.782198] env[62914]: DEBUG oslo_concurrency.lockutils [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "f1c5c877-9c29-4331-84f8-23b3c6b53725" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1392.782316] env[62914]: DEBUG oslo_concurrency.lockutils [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "f1c5c877-9c29-4331-84f8-23b3c6b53725-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1392.782490] env[62914]: DEBUG oslo_concurrency.lockutils [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "f1c5c877-9c29-4331-84f8-23b3c6b53725-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1392.782663] env[62914]: DEBUG oslo_concurrency.lockutils [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "f1c5c877-9c29-4331-84f8-23b3c6b53725-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1392.784903] env[62914]: INFO nova.compute.manager [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Terminating instance [ 1392.786753] env[62914]: DEBUG nova.compute.manager [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Start destroying the instance on the hypervisor. {{(pid=62914) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1392.786952] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Destroying instance {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1392.787797] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6dde64e-8ba1-43a8-a73b-924801d4281a {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.795479] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Powering off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1392.795728] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c78a41b-e692-4801-b022-23392522bbf8 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.801984] env[62914]: DEBUG oslo_vmware.api [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1392.801984] env[62914]: value = "task-1353099" [ 1392.801984] env[62914]: _type = "Task" [ 1392.801984] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1392.809707] env[62914]: DEBUG oslo_vmware.api [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353099, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1393.312752] env[62914]: DEBUG oslo_vmware.api [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353099, 'name': PowerOffVM_Task, 'duration_secs': 0.197886} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1393.313071] env[62914]: DEBUG nova.virt.vmwareapi.vm_util [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Powered off the VM {{(pid=62914) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1393.313283] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Unregistering the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1393.313543] env[62914]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c3e16b91-d7bb-45c0-90e0-5795b4f4f70c {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.383604] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Unregistered the VM {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1393.383862] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Deleting contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1393.384067] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleting the datastore file [datastore2] f1c5c877-9c29-4331-84f8-23b3c6b53725 {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1393.384340] env[62914]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f309369-a103-47bb-b016-4b7b42a96aa6 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.391606] env[62914]: DEBUG oslo_vmware.api [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for the task: (returnval){ [ 1393.391606] env[62914]: value = "task-1353101" [ 1393.391606] env[62914]: _type = "Task" [ 1393.391606] env[62914]: } to complete. {{(pid=62914) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1393.399161] env[62914]: DEBUG oslo_vmware.api [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353101, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1393.901908] env[62914]: DEBUG oslo_vmware.api [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Task: {'id': task-1353101, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146279} completed successfully. {{(pid=62914) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1393.902365] env[62914]: DEBUG nova.virt.vmwareapi.ds_util [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleted the datastore file {{(pid=62914) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1393.902365] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Deleted contents of the VM from datastore datastore2 {{(pid=62914) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1393.902524] env[62914]: DEBUG nova.virt.vmwareapi.vmops [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Instance destroyed {{(pid=62914) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1393.902700] env[62914]: INFO nova.compute.manager [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1393.902938] env[62914]: DEBUG oslo.service.loopingcall [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62914) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1393.903156] env[62914]: DEBUG nova.compute.manager [-] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Deallocating network for instance {{(pid=62914) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1393.903248] env[62914]: DEBUG nova.network.neutron [-] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] deallocate_for_instance() {{(pid=62914) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1394.132451] env[62914]: DEBUG nova.compute.manager [req-52c538f4-99ac-49ac-9b7b-0af793232528 req-2c151292-f8b0-49fe-89b6-6b6d53e86a11 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Received event network-vif-deleted-e4f0d675-9d9c-4f76-bc5b-10ce73f445bf {{(pid=62914) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1394.132641] env[62914]: INFO nova.compute.manager [req-52c538f4-99ac-49ac-9b7b-0af793232528 req-2c151292-f8b0-49fe-89b6-6b6d53e86a11 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Neutron deleted interface e4f0d675-9d9c-4f76-bc5b-10ce73f445bf; detaching it from the instance and deleting it from the info cache [ 1394.132825] env[62914]: DEBUG nova.network.neutron [req-52c538f4-99ac-49ac-9b7b-0af793232528 req-2c151292-f8b0-49fe-89b6-6b6d53e86a11 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1394.608436] env[62914]: DEBUG nova.network.neutron [-] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Updating instance_info_cache with network_info: [] {{(pid=62914) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1394.635824] env[62914]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef5d176b-e684-48ea-9082-15ec83899db1 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.644773] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd32a8f-51fe-4d5f-b6dd-7159d8e8227b {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.665846] env[62914]: DEBUG nova.compute.manager [req-52c538f4-99ac-49ac-9b7b-0af793232528 req-2c151292-f8b0-49fe-89b6-6b6d53e86a11 service nova] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Detach interface failed, port_id=e4f0d675-9d9c-4f76-bc5b-10ce73f445bf, reason: Instance f1c5c877-9c29-4331-84f8-23b3c6b53725 could not be found. {{(pid=62914) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1395.110871] env[62914]: INFO nova.compute.manager [-] [instance: f1c5c877-9c29-4331-84f8-23b3c6b53725] Took 1.21 seconds to deallocate network for instance. [ 1395.617820] env[62914]: DEBUG oslo_concurrency.lockutils [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1395.618140] env[62914]: DEBUG oslo_concurrency.lockutils [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1395.618372] env[62914]: DEBUG nova.objects.instance [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lazy-loading 'resources' on Instance uuid f1c5c877-9c29-4331-84f8-23b3c6b53725 {{(pid=62914) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1396.153855] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f0b874-4cba-43c1-98ce-c6842dc302bf {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.160999] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34dddb7e-7793-4605-951c-1b7d585602a7 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.189568] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de739882-8262-43e6-96ef-15f00d6b3f9f {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.196135] env[62914]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa1df13-4c5f-4d76-8716-7ea74baafdb5 {{(pid=62914) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.208624] env[62914]: DEBUG nova.compute.provider_tree [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed in ProviderTree for provider: 8e1b3185-5ae3-4780-aee8-c507b157ad13 {{(pid=62914) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1396.711428] env[62914]: DEBUG nova.scheduler.client.report [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Inventory has not changed for provider 8e1b3185-5ae3-4780-aee8-c507b157ad13 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62914) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1397.216561] env[62914]: DEBUG oslo_concurrency.lockutils [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.598s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1397.234496] env[62914]: INFO nova.scheduler.client.report [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Deleted allocations for instance f1c5c877-9c29-4331-84f8-23b3c6b53725 [ 1397.742950] env[62914]: DEBUG oslo_concurrency.lockutils [None req-524d9c29-55ea-4315-bec9-ff7e2aafe909 tempest-ServerActionsTestOtherA-761154296 tempest-ServerActionsTestOtherA-761154296-project-member] Lock "f1c5c877-9c29-4331-84f8-23b3c6b53725" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.961s {{(pid=62914) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}